ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible-playbook python version = 2.7.5 (default, Nov 14 2023, 16:14:06) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)] Using /etc/ansible/ansible.cfg as config file [WARNING]: running playbook inside collection fedora.linux_system_roles Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_resize.yml ***************************************************** 1 plays in /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml PLAY [Test resize] ************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:2 Saturday 17 August 2024 19:48:48 -0400 (0:00:00.027) 0:00:00.027 ******* ok: [managed_node2] META: ran handlers TASK [Run the role] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:19 Saturday 17 August 2024 19:48:49 -0400 (0:00:00.942) 0:00:00.969 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:48:49 -0400 (0:00:00.041) 0:00:01.011 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:48:49 -0400 (0:00:00.036) 0:00:01.048 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:48:49 -0400 (0:00:00.051) 0:00:01.099 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:48:49 -0400 (0:00:00.144) 0:00:01.243 ******* ok: [managed_node2] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:48:50 -0400 (0:00:00.413) 0:00:01.657 ******* ok: [managed_node2] => { "ansible_facts": { "__storage_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:48:50 -0400 (0:00:00.053) 0:00:01.710 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:48:50 -0400 (0:00:00.025) 0:00:01.736 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:48:50 -0400 (0:00:00.024) 0:00:01.760 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:48:50 -0400 (0:00:00.086) 0:00:01.846 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "python-enum34-1.0.4-1.el7.noarch providing python-enum34 is already installed", "1:python2-blivet3-3.1.3-3.el7.noarch providing python-blivet3 is already installed", "libblockdev-crypto-2.18-5.el7.x86_64 providing libblockdev-crypto is already installed", "libblockdev-dm-2.18-5.el7.x86_64 providing libblockdev-dm is already installed", "libblockdev-lvm-2.18-5.el7.x86_64 providing libblockdev-lvm is already installed", "libblockdev-mdraid-2.18-5.el7.x86_64 providing libblockdev-mdraid is already installed", "libblockdev-swap-2.18-5.el7.x86_64 providing libblockdev-swap is already installed", "libblockdev-2.18-5.el7.x86_64 providing libblockdev is already installed" ] } lsrpackages: libblockdev libblockdev-crypto libblockdev-dm libblockdev-lvm libblockdev-mdraid libblockdev-swap python-blivet3 python-enum34 TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:48:51 -0400 (0:00:01.358) 0:00:03.205 ******* ok: [managed_node2] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:48:51 -0400 (0:00:00.050) 0:00:03.255 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:48:51 -0400 (0:00:00.049) 0:00:03.305 ******* ok: [managed_node2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:48:52 -0400 (0:00:00.655) 0:00:03.960 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Check if the COPR support packages should be installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Saturday 17 August 2024 19:48:52 -0400 (0:00:00.087) 0:00:04.047 ******* TASK [fedora.linux_system_roles.storage : Make sure COPR support packages are present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Saturday 17 August 2024 19:48:52 -0400 (0:00:00.024) 0:00:04.072 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable COPRs] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:19 Saturday 17 August 2024 19:48:52 -0400 (0:00:00.025) 0:00:04.097 ******* TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:48:52 -0400 (0:00:00.022) 0:00:04.120 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "kpartx-0.4.9-136.el7_9.x86_64 providing kpartx is already installed" ] } lsrpackages: kpartx TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:48:53 -0400 (0:00:00.483) 0:00:04.603 ******* ok: [managed_node2] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "brandbot.service": { "name": "brandbot.service", "source": "systemd", "state": "inactive", "status": "static" }, "chrony-dnssrv@.service": { "name": "chrony-dnssrv@.service", "source": "systemd", "state": "unknown", "status": "static" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "console-shell.service": { "name": "console-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.import1.service": { "name": "dbus-org.freedesktop.import1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "static" }, "dbus-org.freedesktop.machine1.service": { "name": "dbus-org.freedesktop.machine1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "static" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "running", "status": "static" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "ebtables.service": { "name": "ebtables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "halt-local.service": { "name": "halt-local.service", "source": "systemd", "state": "inactive", "status": "static" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "iprdump.service": { "name": "iprdump.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprinit.service": { "name": "iprinit.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "iprupdate.service": { "name": "iprupdate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-lvmetad.service": { "name": "lvm2-lvmetad.service", "source": "systemd", "state": "running", "status": "static" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "messagebus.service": { "name": "messagebus.service", "source": "systemd", "state": "active", "status": "static" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "netconsole": { "name": "netconsole", "source": "sysv", "state": "stopped", "status": "disabled" }, "network": { "name": "network", "source": "sysv", "state": "running", "status": "enabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-config.service": { "name": "nfs-config.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-idmap.service": { "name": "nfs-idmap.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-lock.service": { "name": "nfs-lock.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-rquotad.service": { "name": "nfs-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-secure.service": { "name": "nfs-secure.service", "source": "systemd", "state": "inactive", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs.service": { "name": "nfs.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfslock.service": { "name": "nfslock.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "postfix.service": { "name": "postfix.service", "source": "systemd", "state": "running", "status": "enabled" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rhel-autorelabel-mark.service": { "name": "rhel-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-autorelabel.service": { "name": "rhel-autorelabel.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-configure.service": { "name": "rhel-configure.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-dmesg.service": { "name": "rhel-dmesg.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-domainname.service": { "name": "rhel-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-import-state.service": { "name": "rhel-import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-loadmodules.service": { "name": "rhel-loadmodules.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rhel-readonly.service": { "name": "rhel-readonly.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-rquotad.service": { "name": "rpc-rquotad.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpcgssd.service": { "name": "rpcgssd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rpcidmapd.service": { "name": "rpcidmapd.service", "source": "systemd", "state": "inactive", "status": "static" }, "rsyncd.service": { "name": "rsyncd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyncd@.service": { "name": "rsyncd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-policy-migrate-local-changes@.service": { "name": "selinux-policy-migrate-local-changes@.service", "source": "systemd", "state": "unknown", "status": "static" }, "selinux-policy-migrate-local-changes@targeted.service": { "name": "selinux-policy-migrate-local-changes@targeted.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "unknown" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "static" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bootchart.service": { "name": "systemd-bootchart.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-importd.service": { "name": "systemd-importd.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-machined.service": { "name": "systemd-machined.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-nspawn@.service": { "name": "systemd-nspawn@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-readahead-collect.service": { "name": "systemd-readahead-collect.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-readahead-done.service": { "name": "systemd-readahead-done.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "systemd-readahead-drop.service": { "name": "systemd-readahead-drop.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "systemd-readahead-replay.service": { "name": "systemd-readahead-replay.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill@.service": { "name": "systemd-rfkill@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-shutdownd.service": { "name": "systemd-shutdownd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "wpa_supplicant.service": { "name": "wpa_supplicant.service", "source": "systemd", "state": "inactive", "status": "disabled" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:48:54 -0400 (0:00:01.013) 0:00:05.617 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:48:54 -0400 (0:00:00.047) 0:00:05.664 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:48:54 -0400 (0:00:00.030) 0:00:05.694 ******* ok: [managed_node2] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:48:54 -0400 (0:00:00.455) 0:00:06.150 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:48:54 -0400 (0:00:00.032) 0:00:06.182 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938484.485284, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1723938484.0662827, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938484.0662827, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:48:54 -0400 (0:00:00.321) 0:00:06.503 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.034) 0:00:06.538 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.028) 0:00:06.567 ******* ok: [managed_node2] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.036) 0:00:06.603 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.035) 0:00:06.638 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.037) 0:00:06.676 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.031) 0:00:06.708 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.031) 0:00:06.739 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.030) 0:00:06.769 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.029) 0:00:06.799 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.031) 0:00:06.831 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.309) 0:00:07.141 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:48:55 -0400 (0:00:00.030) 0:00:07.171 ******* ok: [managed_node2] TASK [Mark tasks to be skipped] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:23 Saturday 17 August 2024 19:48:56 -0400 (0:00:00.695) 0:00:07.867 ******* ok: [managed_node2] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Get unused disks] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:30 Saturday 17 August 2024 19:48:56 -0400 (0:00:00.040) 0:00:07.908 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml for managed_node2 TASK [Ensure test packages] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:2 Saturday 17 August 2024 19:48:56 -0400 (0:00:00.054) 0:00:07.962 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "util-linux-2.23.2-65.el7_9.1.x86_64 providing util-linux is already installed" ] } lsrpackages: util-linux TASK [Find unused disks in the system] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:11 Saturday 17 August 2024 19:48:56 -0400 (0:00:00.501) 0:00:08.463 ******* ok: [managed_node2] => { "changed": false, "disks": [ "sda" ], "info": [ "Line: NAME=\"/dev/sda\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdb\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdc\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdd\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sde\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdf\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdg\" TYPE=\"disk\" SIZE=\"1099511627776\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdh\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/sdi\" TYPE=\"disk\" SIZE=\"10737418240\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/xvda\" TYPE=\"disk\" SIZE=\"268435456000\" FSTYPE=\"\" LOG-SEC=\"512\"", "Line: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", "Line type [part] is not disk: NAME=\"/dev/xvda1\" TYPE=\"part\" SIZE=\"268434390528\" FSTYPE=\"ext4\" LOG-SEC=\"512\"", "filename [xvda1] is a partition", "Disk [/dev/xvda] attrs [{'fstype': '', 'type': 'disk', 'ssize': '512', 'size': '268435456000'}] has partitions" ] } TASK [Debug why there are no unused disks] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:20 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.440) 0:00:08.904 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:29 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.032) 0:00:08.937 ******* ok: [managed_node2] => { "ansible_facts": { "unused_disks": [ "sda" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:34 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.038) 0:00:08.975 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/get_unused_disk.yml:39 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.033) 0:00:09.008 ******* ok: [managed_node2] => { "unused_disks": [ "sda" ] } TASK [Create one LVM logical volume under one volume group with size 5g] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:38 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.037) 0:00:09.045 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.064) 0:00:09.110 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.050) 0:00:09.160 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.037) 0:00:09.198 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.081) 0:00:09.279 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.032) 0:00:09.312 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.032) 0:00:09.344 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.032) 0:00:09.377 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.032) 0:00:09.410 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:48:57 -0400 (0:00:00.077) 0:00:09.487 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.036) 0:00:09.524 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.048) 0:00:09.572 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.035) 0:00:09.607 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.035) 0:00:09.643 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.033) 0:00:09.676 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.033) 0:00:09.709 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.032) 0:00:09.742 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.047) 0:00:09.789 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:48:58 -0400 (0:00:00.030) 0:00:09.820 ******* changed: [managed_node2] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:49:04 -0400 (0:00:05.832) 0:00:15.652 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:49:04 -0400 (0:00:00.040) 0:00:15.693 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938484.485284, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1723938484.0662827, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938484.0662827, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:49:04 -0400 (0:00:00.332) 0:00:16.026 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:49:04 -0400 (0:00:00.414) 0:00:16.440 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:49:04 -0400 (0:00:00.030) 0:00:16.471 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:49:05 -0400 (0:00:00.045) 0:00:16.516 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:49:05 -0400 (0:00:00.042) 0:00:16.558 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:49:05 -0400 (0:00:00.037) 0:00:16.596 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:49:05 -0400 (0:00:00.032) 0:00:16.628 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:49:05 -0400 (0:00:00.737) 0:00:17.365 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:49:06 -0400 (0:00:00.444) 0:00:17.810 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:49:06 -0400 (0:00:00.043) 0:00:17.853 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:49:06 -0400 (0:00:00.445) 0:00:18.299 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:49:07 -0400 (0:00:00.315) 0:00:18.615 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:49:07 -0400 (0:00:00.032) 0:00:18.647 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:55 Saturday 17 August 2024 19:49:08 -0400 (0:00:01.694) 0:00:20.342 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:49:08 -0400 (0:00:00.060) 0:00:20.402 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:49:08 -0400 (0:00:00.043) 0:00:20.446 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:49:08 -0400 (0:00:00.032) 0:00:20.479 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "d8157a2a-4f24-4363-8720-5f40504eed14" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "UO145z-7Xnc-7FjF-yNDh-aa4Z-kAQq-IVwMEv" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:49:09 -0400 (0:00:00.448) 0:00:20.928 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002633", "end": "2024-08-17 19:49:09.812473", "rc": 0, "start": "2024-08-17 19:49:09.809840" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:49:09 -0400 (0:00:00.455) 0:00:21.383 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002570", "end": "2024-08-17 19:49:10.123552", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:49:10.120982" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.308) 0:00:21.692 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.074) 0:00:21.767 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.033) 0:00:21.801 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.018719", "end": "2024-08-17 19:49:10.567698", "rc": 0, "start": "2024-08-17 19:49:10.548979" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.337) 0:00:22.138 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.046) 0:00:22.184 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.081) 0:00:22.266 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:49:10 -0400 (0:00:00.044) 0:00:22.310 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.459) 0:00:22.769 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.039) 0:00:22.809 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.041) 0:00:22.850 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.044) 0:00:22.894 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.055) 0:00:22.949 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.040) 0:00:22.990 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.034) 0:00:23.024 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.048) 0:00:23.073 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.233) 0:00:23.306 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.039) 0:00:23.346 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.073) 0:00:23.419 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.035) 0:00:23.455 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:49:11 -0400 (0:00:00.034) 0:00:23.490 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:23.523 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:23.556 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.034) 0:00:23.591 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.034) 0:00:23.625 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.035) 0:00:23.661 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:23.695 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:23.729 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:23.762 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.034) 0:00:23.797 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.075) 0:00:23.873 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.084) 0:00:23.957 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.036) 0:00:23.993 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.035) 0:00:24.029 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:24.062 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:24.096 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.034) 0:00:24.130 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:24.163 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.032) 0:00:24.196 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.077) 0:00:24.273 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.072) 0:00:24.345 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.035) 0:00:24.381 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.034) 0:00:24.415 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.033) 0:00:24.449 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:49:12 -0400 (0:00:00.032) 0:00:24.481 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.079) 0:00:24.561 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.040) 0:00:24.601 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.039) 0:00:24.640 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.065) 0:00:24.705 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.038) 0:00:24.743 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.042) 0:00:24.786 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.033) 0:00:24.820 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.032) 0:00:24.853 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.033) 0:00:24.886 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.033) 0:00:24.920 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.033) 0:00:24.953 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.131) 0:00:25.085 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.077) 0:00:25.163 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.037) 0:00:25.200 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.035) 0:00:25.235 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.034) 0:00:25.270 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.034) 0:00:25.304 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.034) 0:00:25.339 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.035) 0:00:25.374 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.037) 0:00:25.412 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:49:13 -0400 (0:00:00.088) 0:00:25.500 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.036) 0:00:25.536 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.034) 0:00:25.571 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.037) 0:00:25.608 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.033) 0:00:25.642 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.034) 0:00:25.676 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.034) 0:00:25.710 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.034) 0:00:25.745 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.071) 0:00:25.816 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.041) 0:00:25.858 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.198) 0:00:26.056 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.039) 0:00:26.096 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.042) 0:00:26.138 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.035) 0:00:26.174 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.041) 0:00:26.215 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.034) 0:00:26.250 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.036) 0:00:26.286 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.035) 0:00:26.322 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.033) 0:00:26.355 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.033) 0:00:26.388 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.032) 0:00:26.421 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:49:14 -0400 (0:00:00.034) 0:00:26.455 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.060) 0:00:26.516 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.041) 0:00:26.557 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.040) 0:00:26.598 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.033) 0:00:26.632 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.038) 0:00:26.671 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.036) 0:00:26.708 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.046) 0:00:26.754 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.048) 0:00:26.803 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938544.0384827, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938544.0384827, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1201943, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938544.0384827, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.316) 0:00:27.120 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.043) 0:00:27.163 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.034) 0:00:27.198 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.041) 0:00:27.239 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.037) 0:00:27.276 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.033) 0:00:27.310 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.087) 0:00:27.398 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:49:15 -0400 (0:00:00.034) 0:00:27.432 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.516) 0:00:27.949 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:27.984 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:28.018 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.053) 0:00:28.071 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.043) 0:00:28.115 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:28.150 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.033) 0:00:28.184 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.033) 0:00:28.217 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.033) 0:00:28.251 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.041) 0:00:28.292 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.043) 0:00:28.335 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:28.370 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:28.404 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:28.439 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.035) 0:00:28.475 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:49:16 -0400 (0:00:00.034) 0:00:28.509 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.036) 0:00:28.546 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.036) 0:00:28.583 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.035) 0:00:28.618 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.036) 0:00:28.654 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.034) 0:00:28.688 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.034) 0:00:28.723 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.037) 0:00:28.760 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.035) 0:00:28.796 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.034) 0:00:28.830 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:49:17 -0400 (0:00:00.487) 0:00:29.318 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.311) 0:00:29.629 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.045) 0:00:29.674 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.038) 0:00:29.713 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.315) 0:00:30.028 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.039) 0:00:30.068 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.037) 0:00:30.106 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.051) 0:00:30.157 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.039) 0:00:30.197 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.035) 0:00:30.232 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.034) 0:00:30.266 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.033) 0:00:30.300 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.034) 0:00:30.334 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.033) 0:00:30.368 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.034) 0:00:30.403 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.037) 0:00:30.441 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.034) 0:00:30.475 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:49:18 -0400 (0:00:00.034) 0:00:30.509 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.033) 0:00:30.543 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.035) 0:00:30.578 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.033) 0:00:30.612 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.036) 0:00:30.649 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.033) 0:00:30.683 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.034) 0:00:30.717 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.038) 0:00:30.756 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.037) 0:00:30.794 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.048) 0:00:30.842 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.017752", "end": "2024-08-17 19:49:19.599675", "rc": 0, "start": "2024-08-17 19:49:19.581923" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.334) 0:00:31.176 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.047) 0:00:31.224 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.045) 0:00:31.270 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.082) 0:00:31.352 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.038) 0:00:31.390 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.038) 0:00:31.429 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.036) 0:00:31.466 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:49:19 -0400 (0:00:00.033) 0:00:31.500 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.031) 0:00:31.532 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume_size to 9g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:58 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.039) 0:00:31.571 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.076) 0:00:31.648 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.056) 0:00:31.705 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.041) 0:00:31.746 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.085) 0:00:31.832 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.034) 0:00:31.866 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.035) 0:00:31.902 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.035) 0:00:31.937 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.037) 0:00:31.975 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.079) 0:00:32.054 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.035) 0:00:32.089 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "9g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.043) 0:00:32.133 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.037) 0:00:32.171 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.045) 0:00:32.216 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.036) 0:00:32.253 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.034) 0:00:32.287 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.033) 0:00:32.321 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.051) 0:00:32.373 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:49:20 -0400 (0:00:00.032) 0:00:32.405 ******* changed: [managed_node2] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:49:25 -0400 (0:00:04.519) 0:00:36.925 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:49:25 -0400 (0:00:00.034) 0:00:36.960 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938546.22649, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "6b95df90798c1adba66c669f2bd76d2d16d86df5", "ctime": 1723938546.22349, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938546.22349, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:49:25 -0400 (0:00:00.325) 0:00:37.286 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.385) 0:00:37.671 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.032) 0:00:37.703 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.046) 0:00:37.750 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.045) 0:00:37.796 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.039) 0:00:37.835 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.033) 0:00:37.869 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:49:26 -0400 (0:00:00.458) 0:00:38.327 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:49:27 -0400 (0:00:00.353) 0:00:38.680 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:49:27 -0400 (0:00:00.045) 0:00:38.726 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:49:27 -0400 (0:00:00.428) 0:00:39.154 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:49:27 -0400 (0:00:00.309) 0:00:39.464 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:49:27 -0400 (0:00:00.032) 0:00:39.496 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:72 Saturday 17 August 2024 19:49:28 -0400 (0:00:00.668) 0:00:40.165 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:49:28 -0400 (0:00:00.063) 0:00:40.228 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:49:28 -0400 (0:00:00.046) 0:00:40.275 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:49:28 -0400 (0:00:00.034) 0:00:40.310 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "9G", "type": "lvm", "uuid": "d8157a2a-4f24-4363-8720-5f40504eed14" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "UO145z-7Xnc-7FjF-yNDh-aa4Z-kAQq-IVwMEv" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:49:29 -0400 (0:00:00.324) 0:00:40.634 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002769", "end": "2024-08-17 19:49:29.369287", "rc": 0, "start": "2024-08-17 19:49:29.366518" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:49:29 -0400 (0:00:00.305) 0:00:40.940 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002409", "end": "2024-08-17 19:49:29.671200", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:49:29.668791" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:49:29 -0400 (0:00:00.299) 0:00:41.240 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:49:29 -0400 (0:00:00.074) 0:00:41.315 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:49:29 -0400 (0:00:00.034) 0:00:41.349 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.017171", "end": "2024-08-17 19:49:30.110689", "rc": 0, "start": "2024-08-17 19:49:30.093518" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.332) 0:00:41.682 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.047) 0:00:41.729 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.079) 0:00:41.809 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.045) 0:00:41.854 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.311) 0:00:42.165 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.040) 0:00:42.206 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.042) 0:00:42.248 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.043) 0:00:42.292 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.040) 0:00:42.333 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.043) 0:00:42.377 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.035) 0:00:42.412 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:49:30 -0400 (0:00:00.057) 0:00:42.470 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.246) 0:00:42.716 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.040) 0:00:42.757 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.071) 0:00:42.829 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:42.864 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:42.900 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.080) 0:00:42.981 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.016 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.052 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.088 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.034) 0:00:43.123 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.158 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.037) 0:00:43.195 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.231 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.267 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.074) 0:00:43.341 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.077) 0:00:43.418 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.035) 0:00:43.454 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:49:31 -0400 (0:00:00.034) 0:00:43.489 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.047) 0:00:43.536 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.035) 0:00:43.572 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.037) 0:00:43.610 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.035) 0:00:43.646 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.034) 0:00:43.680 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.074) 0:00:43.754 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.078) 0:00:43.833 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.034) 0:00:43.868 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.035) 0:00:43.903 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.034) 0:00:43.937 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.037) 0:00:43.974 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.081) 0:00:44.056 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.040) 0:00:44.097 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.038) 0:00:44.136 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.067) 0:00:44.203 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.039) 0:00:44.242 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.041) 0:00:44.284 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.033) 0:00:44.318 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.035) 0:00:44.353 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.034) 0:00:44.388 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.034) 0:00:44.422 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:49:32 -0400 (0:00:00.033) 0:00:44.456 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.089) 0:00:44.545 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.081) 0:00:44.626 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.035) 0:00:44.662 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.034) 0:00:44.697 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.036) 0:00:44.733 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.035) 0:00:44.769 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.034) 0:00:44.803 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.034) 0:00:44.838 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.035) 0:00:44.873 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.088) 0:00:44.961 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.035) 0:00:44.997 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.034) 0:00:45.031 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.034) 0:00:45.066 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.033) 0:00:45.099 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.080) 0:00:45.180 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.036) 0:00:45.217 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.035) 0:00:45.252 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.069) 0:00:45.322 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:49:33 -0400 (0:00:00.042) 0:00:45.364 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.178) 0:00:45.543 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.053) 0:00:45.597 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.043) 0:00:45.640 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.035) 0:00:45.676 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.042) 0:00:45.719 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.038) 0:00:45.757 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.035) 0:00:45.793 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.034) 0:00:45.828 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.035) 0:00:45.863 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.034) 0:00:45.898 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.034) 0:00:45.932 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.037) 0:00:45.970 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.061) 0:00:46.032 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.043) 0:00:46.075 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.042) 0:00:46.118 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.033) 0:00:46.152 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.043) 0:00:46.195 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.036) 0:00:46.231 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.047) 0:00:46.279 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:49:34 -0400 (0:00:00.050) 0:00:46.330 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938565.3165536, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938565.3165536, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1201943, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938565.3165536, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.315) 0:00:46.645 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.044) 0:00:46.690 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.035) 0:00:46.725 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.042) 0:00:46.767 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.038) 0:00:46.806 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.036) 0:00:46.843 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.041) 0:00:46.884 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.034) 0:00:46.919 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.507) 0:00:47.426 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.035) 0:00:47.462 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:49:35 -0400 (0:00:00.035) 0:00:47.498 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.047) 0:00:47.545 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.035) 0:00:47.581 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.034) 0:00:47.615 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.047) 0:00:47.663 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.034) 0:00:47.697 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.034) 0:00:47.732 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.045) 0:00:47.778 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.042) 0:00:47.820 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.035) 0:00:47.855 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.034) 0:00:47.890 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.034) 0:00:47.924 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.036) 0:00:47.960 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.038) 0:00:47.999 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.035) 0:00:48.034 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.035) 0:00:48.070 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.034) 0:00:48.105 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.033) 0:00:48.138 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.033) 0:00:48.172 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.083) 0:00:48.255 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.035) 0:00:48.291 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.035) 0:00:48.326 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:49:36 -0400 (0:00:00.036) 0:00:48.363 ******* ok: [managed_node2] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.326) 0:00:48.689 ******* ok: [managed_node2] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.311) 0:00:49.000 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "9663676416" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.046) 0:00:49.047 ******* ok: [managed_node2] => { "storage_test_expected_size": "9663676416" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.038) 0:00:49.085 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.311) 0:00:49.396 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.040) 0:00:49.436 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:49:37 -0400 (0:00:00.039) 0:00:49.475 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.039) 0:00:49.515 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.039) 0:00:49.555 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.034) 0:00:49.589 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.037) 0:00:49.627 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.035) 0:00:49.662 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.034) 0:00:49.697 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.046) 0:00:49.743 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.034) 0:00:49.778 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.034) 0:00:49.813 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.036) 0:00:49.849 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.033) 0:00:49.883 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.035) 0:00:49.918 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.034) 0:00:49.952 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.033) 0:00:49.986 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.035) 0:00:50.022 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.038) 0:00:50.060 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.034) 0:00:50.095 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 9663676416, "changed": false, "failed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.039) 0:00:50.135 ******* ok: [managed_node2] => { "storage_test_expected_size": "9663676416" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.038) 0:00:50.173 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:49:38 -0400 (0:00:00.047) 0:00:50.220 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.017574", "end": "2024-08-17 19:49:38.981726", "rc": 0, "start": "2024-08-17 19:49:38.964152" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.332) 0:00:50.553 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.043) 0:00:50.597 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.045) 0:00:50.642 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.039) 0:00:50.682 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.038) 0:00:50.721 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.049) 0:00:50.770 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.038) 0:00:50.808 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.034) 0:00:50.843 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.031) 0:00:50.875 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 5g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:75 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.038) 0:00:50.914 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.079) 0:00:50.993 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.053) 0:00:51.047 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.044) 0:00:51.092 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.083) 0:00:51.175 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.033) 0:00:51.209 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.034) 0:00:51.243 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.035) 0:00:51.279 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.038) 0:00:51.317 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.124) 0:00:51.442 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:49:39 -0400 (0:00:00.035) 0:00:51.477 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.042) 0:00:51.520 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.040) 0:00:51.561 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.035) 0:00:51.596 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.035) 0:00:51.632 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.035) 0:00:51.667 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.035) 0:00:51.703 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.050) 0:00:51.754 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:49:40 -0400 (0:00:00.047) 0:00:51.801 ******* changed: [managed_node2] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:49:44 -0400 (0:00:04.136) 0:00:55.938 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:49:44 -0400 (0:00:00.034) 0:00:55.973 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938546.22649, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "6b95df90798c1adba66c669f2bd76d2d16d86df5", "ctime": 1723938546.22349, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938546.22349, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:49:44 -0400 (0:00:00.315) 0:00:56.288 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.316) 0:00:56.605 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.032) 0:00:56.637 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.047) 0:00:56.685 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.044) 0:00:56.730 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.039) 0:00:56.769 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.034) 0:00:56.803 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:49:45 -0400 (0:00:00.442) 0:00:57.246 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:49:46 -0400 (0:00:00.344) 0:00:57.591 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:49:46 -0400 (0:00:00.045) 0:00:57.636 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:49:46 -0400 (0:00:00.450) 0:00:58.087 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:49:46 -0400 (0:00:00.336) 0:00:58.423 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:49:46 -0400 (0:00:00.033) 0:00:58.456 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:89 Saturday 17 August 2024 19:49:47 -0400 (0:00:00.749) 0:00:59.206 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:49:47 -0400 (0:00:00.068) 0:00:59.275 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:49:47 -0400 (0:00:00.047) 0:00:59.322 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:49:47 -0400 (0:00:00.075) 0:00:59.397 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "d8157a2a-4f24-4363-8720-5f40504eed14" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "UO145z-7Xnc-7FjF-yNDh-aa4Z-kAQq-IVwMEv" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:49:48 -0400 (0:00:00.329) 0:00:59.727 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002669", "end": "2024-08-17 19:49:48.462689", "rc": 0, "start": "2024-08-17 19:49:48.460020" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:49:48 -0400 (0:00:00.309) 0:01:00.036 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002554", "end": "2024-08-17 19:49:48.777944", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:49:48.775390" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:49:48 -0400 (0:00:00.310) 0:01:00.347 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:49:48 -0400 (0:00:00.075) 0:01:00.423 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:49:48 -0400 (0:00:00.036) 0:01:00.459 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.019185", "end": "2024-08-17 19:49:49.215709", "rc": 0, "start": "2024-08-17 19:49:49.196524" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.327) 0:01:00.786 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.046) 0:01:00.833 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.078) 0:01:00.911 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.045) 0:01:00.957 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.313) 0:01:01.270 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.040) 0:01:01.311 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.042) 0:01:01.353 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.045) 0:01:01.399 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.041) 0:01:01.440 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:49:49 -0400 (0:00:00.043) 0:01:01.483 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.034) 0:01:01.518 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.050) 0:01:01.568 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.233) 0:01:01.801 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.040) 0:01:01.842 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.071) 0:01:01.914 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.035) 0:01:01.950 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.037) 0:01:01.987 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.037) 0:01:02.024 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.047) 0:01:02.071 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.035) 0:01:02.107 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.035) 0:01:02.142 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.034) 0:01:02.177 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.037) 0:01:02.215 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.035) 0:01:02.250 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.035) 0:01:02.285 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.037) 0:01:02.322 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.073) 0:01:02.396 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:49:50 -0400 (0:00:00.079) 0:01:02.476 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:02.511 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:02.547 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:02.582 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.077) 0:01:02.660 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:02.696 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.034) 0:01:02.730 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.034) 0:01:02.764 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.076) 0:01:02.841 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.073) 0:01:02.914 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:02.949 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:02.985 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.037) 0:01:03.022 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.039) 0:01:03.062 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.088) 0:01:03.150 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.040) 0:01:03.191 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.042) 0:01:03.234 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.067) 0:01:03.301 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.040) 0:01:03.341 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.042) 0:01:03.383 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.034) 0:01:03.418 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.036) 0:01:03.454 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:49:51 -0400 (0:00:00.035) 0:01:03.489 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:03.525 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:03.560 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.083) 0:01:03.643 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.074) 0:01:03.718 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.034) 0:01:03.753 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.034) 0:01:03.787 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.036) 0:01:03.824 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:03.859 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:03.895 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:03.930 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.033) 0:01:03.964 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.088) 0:01:04.052 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.036) 0:01:04.089 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.046) 0:01:04.135 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.034) 0:01:04.170 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.036) 0:01:04.206 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.034) 0:01:04.241 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:04.276 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.035) 0:01:04.311 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.068) 0:01:04.380 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:49:52 -0400 (0:00:00.045) 0:01:04.425 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.172) 0:01:04.598 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.040) 0:01:04.638 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.044) 0:01:04.683 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:04.718 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.095) 0:01:04.814 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:04.850 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:04.886 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:04.921 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:04.956 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.036) 0:01:04.993 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.037) 0:01:05.030 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:05.066 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.062) 0:01:05.129 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.052) 0:01:05.182 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.042) 0:01:05.224 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.034) 0:01:05.259 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.041) 0:01:05.301 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.035) 0:01:05.336 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.047) 0:01:05.384 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:49:53 -0400 (0:00:00.049) 0:01:05.434 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938584.304617, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938584.304617, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1201943, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938584.304617, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.314) 0:01:05.748 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.045) 0:01:05.794 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.035) 0:01:05.829 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.043) 0:01:05.872 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.041) 0:01:05.914 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.035) 0:01:05.949 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.041) 0:01:05.990 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:49:54 -0400 (0:00:00.034) 0:01:06.025 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.503) 0:01:06.529 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.563 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.036) 0:01:06.600 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.046) 0:01:06.646 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.681 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.715 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.749 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.783 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.037) 0:01:06.821 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.043) 0:01:06.865 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.041) 0:01:06.906 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.941 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:06.975 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:07.010 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.038) 0:01:07.049 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.035) 0:01:07.084 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.035) 0:01:07.120 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:07.155 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:07.189 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.049) 0:01:07.239 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.038) 0:01:07.277 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.035) 0:01:07.313 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.035) 0:01:07.348 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:07.383 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:49:55 -0400 (0:00:00.034) 0:01:07.418 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:49:56 -0400 (0:00:00.325) 0:01:07.744 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:49:56 -0400 (0:00:00.337) 0:01:08.082 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:49:56 -0400 (0:00:00.046) 0:01:08.128 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:49:56 -0400 (0:00:00.038) 0:01:08.167 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:49:56 -0400 (0:00:00.323) 0:01:08.491 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.040) 0:01:08.532 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.040) 0:01:08.572 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.040) 0:01:08.613 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.042) 0:01:08.655 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.078) 0:01:08.734 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.035) 0:01:08.770 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:08.805 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:08.839 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.035) 0:01:08.874 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:08.909 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.037) 0:01:08.946 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:08.981 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.035) 0:01:09.017 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:09.051 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.035) 0:01:09.086 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:09.121 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.036) 0:01:09.158 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.034) 0:01:09.193 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.033) 0:01:09.227 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.043) 0:01:09.271 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.044) 0:01:09.316 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:49:57 -0400 (0:00:00.046) 0:01:09.362 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.018570", "end": "2024-08-17 19:49:58.126657", "rc": 0, "start": "2024-08-17 19:49:58.108087" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.334) 0:01:09.696 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.043) 0:01:09.740 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.045) 0:01:09.786 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.037) 0:01:09.823 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.038) 0:01:09.862 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.036) 0:01:09.899 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.036) 0:01:09.936 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.034) 0:01:09.970 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.034) 0:01:10.005 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Test for correct handling of too-large volume size] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:92 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.033) 0:01:10.039 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed_node2 TASK [Store global variable value copy] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.072) 0:01:10.111 ******* ok: [managed_node2] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": false, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.041) 0:01:10.152 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.061) 0:01:10.214 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.055) 0:01:10.270 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.048) 0:01:10.319 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.091) 0:01:10.410 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.035) 0:01:10.446 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:49:58 -0400 (0:00:00.035) 0:01:10.481 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.035) 0:01:10.516 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.034) 0:01:10.551 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.084) 0:01:10.635 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.035) 0:01:10.671 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "12884901888.0" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.049) 0:01:10.720 ******* ok: [managed_node2] => { "storage_volumes": [] } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.048) 0:01:10.768 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.037) 0:01:10.805 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.034) 0:01:10.840 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.034) 0:01:10.875 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.035) 0:01:10.910 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.050) 0:01:10.960 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:49:59 -0400 (0:00:00.031) 0:01:10.992 ******* fatal: [managed_node2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: volume 'test1' cannot be resized to '12 GiB' TASK [fedora.linux_system_roles.storage : Failed message] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109 Saturday 17 August 2024 19:50:03 -0400 (0:00:04.135) 0:01:15.127 ******* fatal: [managed_node2]: FAILED! => { "changed": false } MSG: {u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': False, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'raid_metadata_version': None, u'mount_device_identifier': u'uuid', u'fs_type': u'ext4', u'mount_options': u'defaults', u'size': u'12884901888.0', u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'encryption': False, u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'deduplication': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'mount_passno': 0, u'mount_user': None, u'raid_spare_count': None, u'name': u'test1', u'cache_mode': None, u'raid_disks': [], u'mount_group': None, u'fs_overwrite_existing': True, u'disks': [u'sda'], u'cached': False, u'thin_pool_size': None, u'thin': False, u'mount_check': 0, u'cache_size': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'lvm', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"volume 'test1' cannot be resized to '12 GiB'"} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.051) 0:01:15.179 ******* TASK [Check that we failed in the role] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.032) 0:01:15.211 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.040) 0:01:15.252 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.049) 0:01:15.301 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Try to create LVM with volume size equal disk's size, resize to 10737418240] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:110 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.040) 0:01:15.341 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.058) 0:01:15.400 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.067) 0:01:15.467 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:50:03 -0400 (0:00:00.042) 0:01:15.509 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.085) 0:01:15.595 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.034) 0:01:15.630 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.036) 0:01:15.667 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.034) 0:01:15.701 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.035) 0:01:15.737 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.083) 0:01:15.820 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.035) 0:01:15.856 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "10737418240" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.047) 0:01:15.904 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.038) 0:01:15.943 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.035) 0:01:15.978 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.035) 0:01:16.014 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.037) 0:01:16.051 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.035) 0:01:16.087 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.051) 0:01:16.138 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:50:04 -0400 (0:00:00.033) 0:01:16.172 ******* changed: [managed_node2] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:50:09 -0400 (0:00:04.486) 0:01:20.659 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:50:09 -0400 (0:00:00.034) 0:01:20.694 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938546.22649, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "6b95df90798c1adba66c669f2bd76d2d16d86df5", "ctime": 1723938546.22349, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938546.22349, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:50:09 -0400 (0:00:00.321) 0:01:21.015 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:50:09 -0400 (0:00:00.317) 0:01:21.333 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:50:09 -0400 (0:00:00.032) 0:01:21.365 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:50:09 -0400 (0:00:00.120) 0:01:21.486 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:50:10 -0400 (0:00:00.045) 0:01:21.531 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:50:10 -0400 (0:00:00.041) 0:01:21.573 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:50:10 -0400 (0:00:00.047) 0:01:21.620 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:50:10 -0400 (0:00:00.447) 0:01:22.067 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:50:10 -0400 (0:00:00.345) 0:01:22.413 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext4', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:50:10 -0400 (0:00:00.045) 0:01:22.458 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:50:11 -0400 (0:00:00.468) 0:01:22.927 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:50:11 -0400 (0:00:00.333) 0:01:23.260 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:50:11 -0400 (0:00:00.033) 0:01:23.293 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:125 Saturday 17 August 2024 19:50:12 -0400 (0:00:00.692) 0:01:23.986 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:50:12 -0400 (0:00:00.058) 0:01:24.044 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:50:12 -0400 (0:00:00.046) 0:01:24.091 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:50:12 -0400 (0:00:00.034) 0:01:24.125 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "10G", "type": "lvm", "uuid": "d8157a2a-4f24-4363-8720-5f40504eed14" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "UO145z-7Xnc-7FjF-yNDh-aa4Z-kAQq-IVwMEv" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:50:12 -0400 (0:00:00.310) 0:01:24.436 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002575", "end": "2024-08-17 19:50:13.179555", "rc": 0, "start": "2024-08-17 19:50:13.176980" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:50:13 -0400 (0:00:00.315) 0:01:24.751 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002459", "end": "2024-08-17 19:50:13.492218", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:50:13.489759" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:50:13 -0400 (0:00:00.310) 0:01:25.062 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:50:13 -0400 (0:00:00.076) 0:01:25.138 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:50:13 -0400 (0:00:00.035) 0:01:25.174 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.018056", "end": "2024-08-17 19:50:13.931660", "rc": 0, "start": "2024-08-17 19:50:13.913604" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:50:13 -0400 (0:00:00.329) 0:01:25.503 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.047) 0:01:25.551 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.078) 0:01:25.629 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.046) 0:01:25.675 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.324) 0:01:26.000 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.040) 0:01:26.041 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.042) 0:01:26.083 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.045) 0:01:26.129 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.041) 0:01:26.170 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.040) 0:01:26.211 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.035) 0:01:26.246 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:50:14 -0400 (0:00:00.049) 0:01:26.296 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.238) 0:01:26.534 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.040) 0:01:26.574 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.073) 0:01:26.647 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.095) 0:01:26.743 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.038) 0:01:26.781 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:26.817 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:26.852 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:26.888 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:26.923 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.038) 0:01:26.962 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:26.998 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.036) 0:01:27.035 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:27.071 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.036) 0:01:27.107 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.076) 0:01:27.184 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.076) 0:01:27.260 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.037) 0:01:27.298 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.037) 0:01:27.335 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:27.371 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:27.406 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:27.442 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:50:15 -0400 (0:00:00.035) 0:01:27.478 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.035) 0:01:27.513 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.077) 0:01:27.591 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.074) 0:01:27.665 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.036) 0:01:27.702 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.034) 0:01:27.737 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.049) 0:01:27.786 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.035) 0:01:27.822 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.082) 0:01:27.905 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.041) 0:01:27.946 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.040) 0:01:27.987 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.068) 0:01:28.055 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.042) 0:01:28.097 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.044) 0:01:28.141 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.035) 0:01:28.177 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.033) 0:01:28.211 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.034) 0:01:28.245 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.034) 0:01:28.280 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.034) 0:01:28.315 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.084) 0:01:28.399 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:50:16 -0400 (0:00:00.076) 0:01:28.476 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.037) 0:01:28.514 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.036) 0:01:28.550 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.035) 0:01:28.586 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.035) 0:01:28.621 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.036) 0:01:28.658 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.035) 0:01:28.693 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.037) 0:01:28.731 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.102) 0:01:28.834 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.034) 0:01:28.868 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.037) 0:01:28.906 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.085) 0:01:28.991 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.036) 0:01:29.028 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.036) 0:01:29.064 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.035) 0:01:29.100 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.035) 0:01:29.135 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.073) 0:01:29.209 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.042) 0:01:29.252 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.176) 0:01:29.429 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:50:17 -0400 (0:00:00.041) 0:01:29.471 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.047) 0:01:29.518 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.036) 0:01:29.555 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.043) 0:01:29.599 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.036) 0:01:29.635 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.036) 0:01:29.671 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.035) 0:01:29.707 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.038) 0:01:29.745 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.035) 0:01:29.781 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.035) 0:01:29.816 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.048) 0:01:29.864 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.060) 0:01:29.925 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.042) 0:01:29.968 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.044) 0:01:30.013 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.035) 0:01:30.048 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.040) 0:01:30.088 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.036) 0:01:30.124 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.047) 0:01:30.172 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:50:18 -0400 (0:00:00.049) 0:01:30.221 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938609.0526993, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938609.0526993, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1201943, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938609.0526993, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.318) 0:01:30.539 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.045) 0:01:30.585 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.037) 0:01:30.623 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.043) 0:01:30.666 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.038) 0:01:30.704 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.034) 0:01:30.739 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.041) 0:01:30.781 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.035) 0:01:30.816 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.501) 0:01:31.317 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.035) 0:01:31.353 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.035) 0:01:31.388 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.047) 0:01:31.436 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.034) 0:01:31.470 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:50:19 -0400 (0:00:00.034) 0:01:31.504 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.034) 0:01:31.538 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.037) 0:01:31.576 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.034) 0:01:31.611 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.043) 0:01:31.654 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.041) 0:01:31.696 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.035) 0:01:31.731 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.034) 0:01:31.766 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.087) 0:01:31.854 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.040) 0:01:31.894 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.043) 0:01:31.937 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.034) 0:01:31.972 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.035) 0:01:32.008 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.035) 0:01:32.043 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.038) 0:01:32.082 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.036) 0:01:32.119 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.037) 0:01:32.156 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.035) 0:01:32.192 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.035) 0:01:32.227 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:50:20 -0400 (0:00:00.036) 0:01:32.263 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.341) 0:01:32.604 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.336) 0:01:32.941 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "10737418240" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.052) 0:01:32.993 ******* ok: [managed_node2] => { "storage_test_expected_size": "10737418240" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.038) 0:01:33.032 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.327) 0:01:33.360 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.040) 0:01:33.400 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.040) 0:01:33.441 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:50:21 -0400 (0:00:00.041) 0:01:33.483 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.040) 0:01:33.523 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.558 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.594 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.034) 0:01:33.628 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.034) 0:01:33.663 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.038) 0:01:33.701 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.737 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.773 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.808 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.036) 0:01:33.845 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.880 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.038) 0:01:33.918 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:33.953 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.051) 0:01:34.004 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.034) 0:01:34.039 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.035) 0:01:34.075 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.040) 0:01:34.115 ******* ok: [managed_node2] => { "storage_test_expected_size": "10737418240" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.041) 0:01:34.156 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:50:22 -0400 (0:00:00.046) 0:01:34.203 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.018702", "end": "2024-08-17 19:50:22.960732", "rc": 0, "start": "2024-08-17 19:50:22.942030" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.328) 0:01:34.532 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.044) 0:01:34.576 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.045) 0:01:34.621 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.038) 0:01:34.659 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.038) 0:01:34.698 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.038) 0:01:34.736 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.038) 0:01:34.774 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.038) 0:01:34.813 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.032) 0:01:34.845 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Test for correct handling of invalid size specification] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:128 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.035) 0:01:34.881 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed_node2 TASK [Store global variable value copy] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.058) 0:01:34.940 ******* ok: [managed_node2] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": false, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.044) 0:01:34.984 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.065) 0:01:35.049 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.055) 0:01:35.105 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.043) 0:01:35.149 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.136) 0:01:35.285 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.034) 0:01:35.320 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.034) 0:01:35.355 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.035) 0:01:35.391 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:50:23 -0400 (0:00:00.035) 0:01:35.426 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.085) 0:01:35.512 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.035) 0:01:35.548 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "xyz GiB" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.045) 0:01:35.593 ******* ok: [managed_node2] => { "storage_volumes": [] } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.044) 0:01:35.638 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.038) 0:01:35.677 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.036) 0:01:35.713 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.036) 0:01:35.749 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.036) 0:01:35.786 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.051) 0:01:35.838 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:50:24 -0400 (0:00:00.032) 0:01:35.871 ******* fatal: [managed_node2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: invalid size specification 'xyz GiB' in pool 'foo' TASK [fedora.linux_system_roles.storage : Failed message] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109 Saturday 17 August 2024 19:50:28 -0400 (0:00:03.723) 0:01:39.595 ******* fatal: [managed_node2]: FAILED! => { "changed": false } MSG: {u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': False, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'thin': False, u'encryption_key_size': None, u'encryption_key': None, u'fs_label': None, u'encryption_luks_version': None, u'raid_stripe_size': None, u'fs_type': u'ext4', u'mount_options': None, u'type': None, u'mount_user': None, u'raid_disks': [], u'size': u'xyz GiB', u'cache_mode': None, u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'mount_group': None, u'encryption': None, u'raid_level': None, u'name': u'test1', u'cached': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'cache_size': None, u'thin_pool_name': None, u'cache_devices': [], u'thin_pool_size': None, u'encryption_cipher': None, u'fs_create_options': None, u'deduplication': None}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'lvm', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"invalid size specification 'xyz GiB' in pool 'foo'"} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.049) 0:01:39.645 ******* TASK [Check that we failed in the role] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.031) 0:01:39.676 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.040) 0:01:39.717 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.049) 0:01:39.767 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Test for correct handling of invalid size specification] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:144 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.036) 0:01:39.803 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml for managed_node2 TASK [Store global variable value copy] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:4 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.064) 0:01:39.867 ******* ok: [managed_node2] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": false, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:10 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.042) 0:01:39.910 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.053) 0:01:39.963 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.058) 0:01:40.022 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.042) 0:01:40.064 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.099) 0:01:40.163 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.034) 0:01:40.198 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.034) 0:01:40.233 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.037) 0:01:40.271 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.035) 0:01:40.306 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.084) 0:01:40.391 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.038) 0:01:40.430 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "none" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:50:28 -0400 (0:00:00.045) 0:01:40.476 ******* ok: [managed_node2] => { "storage_volumes": [] } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.044) 0:01:40.520 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.035) 0:01:40.556 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.035) 0:01:40.591 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.035) 0:01:40.627 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.038) 0:01:40.666 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.098) 0:01:40.764 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:50:29 -0400 (0:00:00.034) 0:01:40.798 ******* fatal: [managed_node2]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: invalid size specification 'none' in pool 'foo' TASK [fedora.linux_system_roles.storage : Failed message] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:109 Saturday 17 August 2024 19:50:33 -0400 (0:00:04.097) 0:01:44.896 ******* fatal: [managed_node2]: FAILED! => { "changed": false } MSG: {u'_ansible_no_log': False, u'crypts': [], u'pools': [], u'leaves': [], u'changed': False, u'actions': [], u'failed': True, u'volumes': [], u'invocation': {u'module_args': {u'packages_only': False, u'disklabel_type': None, u'diskvolume_mkfs_option_map': {u'ext4': u'-F', u'ext3': u'-F', u'ext2': u'-F'}, u'safe_mode': False, u'pools': [{u'raid_metadata_version': None, u'encryption_key_size': None, u'encryption_key': None, u'encryption_luks_version': None, u'encryption_tang_url': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_tang_thumbprint': None, u'name': u'foo', u'encryption_password': None, u'encryption': False, u'disks': [u'sda'], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [{u'thin': False, u'encryption_key_size': None, u'encryption_key': None, u'fs_label': None, u'encryption_luks_version': None, u'raid_stripe_size': None, u'fs_type': u'ext4', u'mount_options': None, u'type': None, u'mount_user': None, u'raid_disks': [], u'size': u'none', u'cache_mode': None, u'mount_point': u'/opt/test1', u'compression': None, u'encryption_password': None, u'mount_group': None, u'encryption': None, u'raid_level': None, u'name': u'test1', u'cached': None, u'state': u'present', u'vdo_pool_size': None, u'mount_mode': None, u'cache_size': None, u'thin_pool_name': None, u'cache_devices': [], u'thin_pool_size': None, u'encryption_cipher': None, u'fs_create_options': None, u'deduplication': None}], u'shared': False, u'encryption_clevis_pin': None, u'type': u'lvm', u'encryption_cipher': None, u'raid_chunk_size': None}], u'volumes': [], u'pool_defaults': {u'raid_metadata_version': None, u'encryption_cipher': None, u'encryption_key': None, u'encryption_luks_version': None, u'raid_spare_count': None, u'grow_to_fill': False, u'encryption_password': None, u'encryption': False, u'disks': [], u'raid_level': None, u'raid_device_count': None, u'state': u'present', u'volumes': [], u'shared': False, u'type': u'lvm', u'encryption_key_size': None, u'raid_chunk_size': None}, u'volume_defaults': {u'raid_metadata_version': None, u'raid_level': None, u'fs_type': u'xfs', u'mount_options': u'defaults', u'size': 0, u'mount_point': u'', u'compression': None, u'encryption_password': None, u'encryption': False, u'mount_device_identifier': u'uuid', u'raid_device_count': None, u'state': u'present', u'vdo_pool_size': None, u'thin_pool_name': None, u'type': u'lvm', u'encryption_key_size': None, u'encryption_cipher': None, u'encryption_key': None, u'fs_label': u'', u'encryption_luks_version': None, u'raid_stripe_size': None, u'cache_size': 0, u'raid_spare_count': None, u'cache_mode': None, u'deduplication': None, u'cached': False, u'fs_overwrite_existing': True, u'disks': [], u'thin_pool_size': None, u'thin': None, u'mount_check': 0, u'mount_passno': 0, u'raid_chunk_size': None, u'cache_devices': [], u'fs_create_options': u''}, u'use_partitions': None}}, u'mounts': [], u'packages': [], u'msg': u"invalid size specification 'none' in pool 'foo'"} TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.049) 0:01:44.945 ******* TASK [Check that we failed in the role] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:23 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.032) 0:01:44.977 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:28 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.040) 0:01:45.018 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-failed.yml:39 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.049) 0:01:45.067 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up] **************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:160 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.036) 0:01:45.104 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.079) 0:01:45.183 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.059) 0:01:45.243 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.050) 0:01:45.293 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.083) 0:01:45.377 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.034) 0:01:45.411 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.034) 0:01:45.446 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:50:33 -0400 (0:00:00.038) 0:01:45.484 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.036) 0:01:45.520 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.080) 0:01:45.601 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.036) 0:01:45.637 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.045) 0:01:45.682 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.039) 0:01:45.721 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.034) 0:01:45.756 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.035) 0:01:45.792 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.034) 0:01:45.826 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.034) 0:01:45.861 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.054) 0:01:45.915 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:50:34 -0400 (0:00:00.032) 0:01:45.948 ******* changed: [managed_node2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:50:38 -0400 (0:00:04.216) 0:01:50.164 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:50:38 -0400 (0:00:00.034) 0:01:50.199 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938546.22649, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "6b95df90798c1adba66c669f2bd76d2d16d86df5", "ctime": 1723938546.22349, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938546.22349, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.331) 0:01:50.530 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.316) 0:01:50.846 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.031) 0:01:50.878 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.046) 0:01:50.924 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.044) 0:01:50.968 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.040) 0:01:51.009 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'state': u'absent', u'fstype': u'ext4', u'path': u'/opt/test1'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "fstype": "ext4", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:50:39 -0400 (0:00:00.330) 0:01:51.340 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:50:40 -0400 (0:00:00.457) 0:01:51.797 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:50:40 -0400 (0:00:00.034) 0:01:51.831 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:50:40 -0400 (0:00:00.032) 0:01:51.864 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:50:40 -0400 (0:00:00.436) 0:01:52.301 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:50:41 -0400 (0:00:00.357) 0:01:52.658 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:50:41 -0400 (0:00:00.033) 0:01:52.692 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:173 Saturday 17 August 2024 19:50:41 -0400 (0:00:00.683) 0:01:53.375 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:50:41 -0400 (0:00:00.071) 0:01:53.446 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:50:41 -0400 (0:00:00.053) 0:01:53.499 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:50:42 -0400 (0:00:00.034) 0:01:53.534 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:50:42 -0400 (0:00:00.320) 0:01:53.855 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002687", "end": "2024-08-17 19:50:42.594750", "rc": 0, "start": "2024-08-17 19:50:42.592063" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:50:42 -0400 (0:00:00.311) 0:01:54.167 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002603", "end": "2024-08-17 19:50:42.901753", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:50:42.899150" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:50:42 -0400 (0:00:00.305) 0:01:54.473 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.083) 0:01:54.556 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.036) 0:01:54.592 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.034) 0:01:54.627 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.034) 0:01:54.662 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.078) 0:01:54.740 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.047) 0:01:54.788 ******* TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.031) 0:01:54.820 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.040) 0:01:54.860 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.042) 0:01:54.903 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.044) 0:01:54.947 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.040) 0:01:54.988 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.043) 0:01:55.031 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.034) 0:01:55.066 ******* TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.030) 0:01:55.096 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.229) 0:01:55.326 ******* TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.031) 0:01:55.357 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.070) 0:01:55.427 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.034) 0:01:55.462 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:50:43 -0400 (0:00:00.036) 0:01:55.498 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.044) 0:01:55.543 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.036) 0:01:55.580 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.033) 0:01:55.614 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.033) 0:01:55.648 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.033) 0:01:55.681 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.036) 0:01:55.717 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.082) 0:01:55.800 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.034) 0:01:55.835 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.034) 0:01:55.869 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.072) 0:01:55.942 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.076) 0:01:56.019 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.035) 0:01:56.054 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.034) 0:01:56.088 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.035) 0:01:56.123 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.033) 0:01:56.157 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.037) 0:01:56.195 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.034) 0:01:56.229 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.033) 0:01:56.263 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.073) 0:01:56.336 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.074) 0:01:56.411 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.034) 0:01:56.445 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:50:44 -0400 (0:00:00.033) 0:01:56.478 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.033) 0:01:56.511 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.034) 0:01:56.546 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.094) 0:01:56.641 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.040) 0:01:56.681 ******* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.031) 0:01:56.713 ******* TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.030) 0:01:56.744 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.037) 0:01:56.781 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.079) 0:01:56.861 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.076) 0:01:56.937 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.037) 0:01:56.974 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.035) 0:01:57.010 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.034) 0:01:57.044 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.034) 0:01:57.079 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.035) 0:01:57.114 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.034) 0:01:57.149 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.037) 0:01:57.186 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.086) 0:01:57.272 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.035) 0:01:57.308 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.034) 0:01:57.342 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.036) 0:01:57.378 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.035) 0:01:57.413 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.033) 0:01:57.447 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:50:45 -0400 (0:00:00.034) 0:01:57.482 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.034) 0:01:57.516 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.072) 0:01:57.588 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.051) 0:01:57.640 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.173) 0:01:57.813 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.039) 0:01:57.853 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.043) 0:01:57.896 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.037) 0:01:57.934 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.084) 0:01:58.019 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.036) 0:01:58.055 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.035) 0:01:58.090 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.035) 0:01:58.126 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.034) 0:01:58.160 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.034) 0:01:58.195 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.036) 0:01:58.232 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.035) 0:01:58.268 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.060) 0:01:58.329 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.034) 0:01:58.363 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.042) 0:01:58.406 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.034) 0:01:58.440 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:50:46 -0400 (0:00:00.043) 0:01:58.484 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.036) 0:01:58.520 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.036) 0:01:58.557 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.036) 0:01:58.594 ******* ok: [managed_node2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.314) 0:01:58.908 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.035) 0:01:58.944 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.044) 0:01:58.989 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.030) 0:01:59.019 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.037) 0:01:59.057 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.036) 0:01:59.093 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.029) 0:01:59.123 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:50:47 -0400 (0:00:00.034) 0:01:59.157 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.508) 0:01:59.666 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.040) 0:01:59.706 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.035) 0:01:59.741 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.030) 0:01:59.771 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:01:59.806 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.036) 0:01:59.842 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.028) 0:01:59.871 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.029) 0:01:59.900 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.029) 0:01:59.930 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.045) 0:01:59.976 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.041) 0:02:00.017 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.052 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.033) 0:02:00.086 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.033) 0:02:00.119 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.154 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.035) 0:02:00.190 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.225 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.260 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.033) 0:02:00.293 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.035) 0:02:00.329 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.363 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.036) 0:02:00.400 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.435 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.035) 0:02:00.470 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:50:48 -0400 (0:00:00.034) 0:02:00.504 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.539 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.033) 0:02:00.572 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.036) 0:02:00.609 ******* ok: [managed_node2] => { "storage_test_expected_size": "10737418240" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.038) 0:02:00.647 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.681 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.046) 0:02:00.728 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.762 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.796 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.036) 0:02:00.833 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.867 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.901 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.033) 0:02:00.934 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:00.969 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:01.003 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.036) 0:02:01.040 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.035) 0:02:01.076 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:01.110 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.035) 0:02:01.146 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.033) 0:02:01.179 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:01.214 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.036) 0:02:01.250 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.034) 0:02:01.285 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.033) 0:02:01.318 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.033) 0:02:01.352 ******* ok: [managed_node2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.038) 0:02:01.391 ******* ok: [managed_node2] => { "storage_test_expected_size": "10737418240" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.038) 0:02:01.430 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.037) 0:02:01.467 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:50:49 -0400 (0:00:00.036) 0:02:01.503 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.035) 0:02:01.539 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.034) 0:02:01.573 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.035) 0:02:01.609 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.033) 0:02:01.643 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.036) 0:02:01.680 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.034) 0:02:01.714 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.048) 0:02:01.762 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.031) 0:02:01.794 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a LVM logical volume with for ext3 FS size 5g] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:178 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.036) 0:02:01.830 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.142) 0:02:01.973 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.052) 0:02:02.025 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.041) 0:02:02.066 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.088) 0:02:02.155 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.035) 0:02:02.190 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.034) 0:02:02.224 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.034) 0:02:02.259 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.034) 0:02:02.293 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.080) 0:02:02.374 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.035) 0:02:02.409 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext3", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.042) 0:02:02.452 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:50:50 -0400 (0:00:00.038) 0:02:02.490 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:50:51 -0400 (0:00:00.037) 0:02:02.528 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:50:51 -0400 (0:00:00.035) 0:02:02.563 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:50:51 -0400 (0:00:00.035) 0:02:02.598 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:50:51 -0400 (0:00:00.034) 0:02:02.633 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:50:51 -0400 (0:00:00.050) 0:02:02.683 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:50:51 -0400 (0:00:00.030) 0:02:02.714 ******* changed: [managed_node2] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:50:58 -0400 (0:00:07.353) 0:02:10.067 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:50:58 -0400 (0:00:00.034) 0:02:10.102 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938640.1818032, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1723938639.7648017, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938639.7648017, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:50:58 -0400 (0:00:00.327) 0:02:10.430 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.319) 0:02:10.749 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.033) 0:02:10.782 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.047) 0:02:10.830 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.043) 0:02:10.873 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.041) 0:02:10.915 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.039) 0:02:10.955 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:50:59 -0400 (0:00:00.456) 0:02:11.411 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext3', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:51:00 -0400 (0:00:00.376) 0:02:11.788 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext3', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:51:00 -0400 (0:00:00.046) 0:02:11.834 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:51:00 -0400 (0:00:00.447) 0:02:12.282 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:51:01 -0400 (0:00:00.317) 0:02:12.600 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:51:01 -0400 (0:00:00.033) 0:02:12.633 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:194 Saturday 17 August 2024 19:51:01 -0400 (0:00:00.695) 0:02:13.329 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:51:01 -0400 (0:00:00.073) 0:02:13.402 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:51:01 -0400 (0:00:00.046) 0:02:13.449 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:51:01 -0400 (0:00:00.037) 0:02:13.486 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext3", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "4a67b7e5-56e4-4a75-9eda-5719be04ec64" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "1Y1RS3-4KcE-61ng-zlkk-uC1e-VKgr-w1g5I3" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:51:02 -0400 (0:00:00.436) 0:02:13.923 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002973", "end": "2024-08-17 19:51:02.697453", "rc": 0, "start": "2024-08-17 19:51:02.694480" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:51:02 -0400 (0:00:00.347) 0:02:14.270 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002880", "end": "2024-08-17 19:51:03.017142", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:51:03.014262" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.320) 0:02:14.591 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.078) 0:02:14.670 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.037) 0:02:14.707 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.017632", "end": "2024-08-17 19:51:03.474677", "rc": 0, "start": "2024-08-17 19:51:03.457045" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.339) 0:02:15.047 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.061) 0:02:15.108 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.082) 0:02:15.190 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:51:03 -0400 (0:00:00.046) 0:02:15.236 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.330) 0:02:15.567 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.043) 0:02:15.610 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.044) 0:02:15.654 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.045) 0:02:15.700 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.044) 0:02:15.745 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.042) 0:02:15.788 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.036) 0:02:15.824 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.051) 0:02:15.876 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.228) 0:02:16.104 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.052) 0:02:16.156 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.074) 0:02:16.231 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.038) 0:02:16.269 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.036) 0:02:16.305 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.035) 0:02:16.341 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.034) 0:02:16.376 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.035) 0:02:16.411 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.035) 0:02:16.446 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:51:04 -0400 (0:00:00.038) 0:02:16.485 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.035) 0:02:16.521 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.036) 0:02:16.557 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.036) 0:02:16.594 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.036) 0:02:16.630 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.076) 0:02:16.707 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.075) 0:02:16.782 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.035) 0:02:16.818 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.035) 0:02:16.853 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.084) 0:02:16.938 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.036) 0:02:16.974 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.036) 0:02:17.010 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.035) 0:02:17.046 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.035) 0:02:17.082 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.091) 0:02:17.174 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.077) 0:02:17.251 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.036) 0:02:17.287 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.037) 0:02:17.325 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.034) 0:02:17.360 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.035) 0:02:17.395 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:51:05 -0400 (0:00:00.080) 0:02:17.476 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.043) 0:02:17.520 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.041) 0:02:17.562 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.068) 0:02:17.630 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.040) 0:02:17.671 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.041) 0:02:17.713 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.037) 0:02:17.751 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.035) 0:02:17.786 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.035) 0:02:17.822 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.036) 0:02:17.859 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.036) 0:02:17.895 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.084) 0:02:17.980 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.077) 0:02:18.058 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.036) 0:02:18.094 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.038) 0:02:18.133 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.036) 0:02:18.169 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.047) 0:02:18.217 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.035) 0:02:18.253 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.035) 0:02:18.288 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.035) 0:02:18.324 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.089) 0:02:18.413 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.034) 0:02:18.448 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:51:06 -0400 (0:00:00.035) 0:02:18.484 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.038) 0:02:18.522 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.035) 0:02:18.558 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.035) 0:02:18.593 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.035) 0:02:18.629 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.035) 0:02:18.664 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.073) 0:02:18.737 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.043) 0:02:18.781 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.176) 0:02:18.957 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.040) 0:02:18.998 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.042) 0:02:19.041 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.097) 0:02:19.138 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.045) 0:02:19.183 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.041) 0:02:19.225 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.042) 0:02:19.267 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.036) 0:02:19.303 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.035) 0:02:19.339 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.038) 0:02:19.378 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.035) 0:02:19.414 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:51:07 -0400 (0:00:00.036) 0:02:19.451 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext3 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.062) 0:02:19.514 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.044) 0:02:19.558 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.043) 0:02:19.602 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.038) 0:02:19.641 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.041) 0:02:19.683 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.037) 0:02:19.720 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.049) 0:02:19.770 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.051) 0:02:19.822 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938658.447864, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938658.447864, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1230158, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938658.447864, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.326) 0:02:20.148 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.047) 0:02:20.196 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.037) 0:02:20.233 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.058) 0:02:20.291 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.040) 0:02:20.331 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.036) 0:02:20.368 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.042) 0:02:20.411 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:51:08 -0400 (0:00:00.036) 0:02:20.447 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.499) 0:02:20.947 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.037) 0:02:20.985 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.036) 0:02:21.021 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.048) 0:02:21.070 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.036) 0:02:21.106 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.035) 0:02:21.142 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.035) 0:02:21.178 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.038) 0:02:21.216 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.036) 0:02:21.253 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.050) 0:02:21.304 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.049) 0:02:21.353 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.035) 0:02:21.389 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.035) 0:02:21.424 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.039) 0:02:21.464 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:51:09 -0400 (0:00:00.036) 0:02:21.501 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.035) 0:02:21.536 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.036) 0:02:21.573 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.035) 0:02:21.609 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.036) 0:02:21.645 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.039) 0:02:21.685 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.037) 0:02:21.722 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.036) 0:02:21.759 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.036) 0:02:21.796 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.036) 0:02:21.832 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.036) 0:02:21.868 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:51:10 -0400 (0:00:00.317) 0:02:22.186 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.327) 0:02:22.513 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.046) 0:02:22.559 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.038) 0:02:22.598 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.329) 0:02:22.928 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.041) 0:02:22.969 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.040) 0:02:23.009 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.040) 0:02:23.050 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.042) 0:02:23.093 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.035) 0:02:23.128 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.036) 0:02:23.164 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.035) 0:02:23.200 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.035) 0:02:23.236 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.035) 0:02:23.271 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.092) 0:02:23.364 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.045) 0:02:23.409 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.036) 0:02:23.446 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:51:11 -0400 (0:00:00.035) 0:02:23.481 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.035) 0:02:23.517 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.035) 0:02:23.553 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.039) 0:02:23.592 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.036) 0:02:23.628 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.036) 0:02:23.665 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.036) 0:02:23.702 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.040) 0:02:23.742 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.038) 0:02:23.781 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.050) 0:02:23.832 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.017263", "end": "2024-08-17 19:51:12.593693", "rc": 0, "start": "2024-08-17 19:51:12.576430" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.335) 0:02:24.167 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.046) 0:02:24.214 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.046) 0:02:24.261 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.040) 0:02:24.301 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.039) 0:02:24.340 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.038) 0:02:24.379 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.052) 0:02:24.432 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.040) 0:02:24.472 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:51:12 -0400 (0:00:00.033) 0:02:24.506 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 9g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:197 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.037) 0:02:24.543 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.102) 0:02:24.645 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.054) 0:02:24.700 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.042) 0:02:24.743 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.087) 0:02:24.830 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.038) 0:02:24.868 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.035) 0:02:24.904 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.036) 0:02:24.940 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.035) 0:02:24.976 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.083) 0:02:25.060 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.037) 0:02:25.098 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext3", "mount_point": "/opt/test1", "name": "test1", "size": "9g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.043) 0:02:25.142 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.039) 0:02:25.181 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.036) 0:02:25.217 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.038) 0:02:25.256 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.036) 0:02:25.292 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.036) 0:02:25.328 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.051) 0:02:25.380 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:51:13 -0400 (0:00:00.032) 0:02:25.413 ******* changed: [managed_node2] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:51:19 -0400 (0:00:05.257) 0:02:30.670 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:51:19 -0400 (0:00:00.034) 0:02:30.705 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938660.20387, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "1ff1203632e7c5aaed867a74eb25885b038df3b8", "ctime": 1723938660.20187, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938660.20187, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:51:19 -0400 (0:00:00.333) 0:02:31.039 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:51:19 -0400 (0:00:00.336) 0:02:31.376 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:51:19 -0400 (0:00:00.033) 0:02:31.410 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:51:19 -0400 (0:00:00.048) 0:02:31.458 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:51:19 -0400 (0:00:00.047) 0:02:31.506 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:51:20 -0400 (0:00:00.041) 0:02:31.547 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:51:20 -0400 (0:00:00.040) 0:02:31.588 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:51:20 -0400 (0:00:00.453) 0:02:32.042 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext3', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:51:20 -0400 (0:00:00.343) 0:02:32.385 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext3', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:51:20 -0400 (0:00:00.045) 0:02:32.431 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:51:21 -0400 (0:00:00.455) 0:02:32.887 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:51:21 -0400 (0:00:00.318) 0:02:33.205 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:51:21 -0400 (0:00:00.033) 0:02:33.239 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:211 Saturday 17 August 2024 19:51:22 -0400 (0:00:00.669) 0:02:33.908 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:51:22 -0400 (0:00:00.076) 0:02:33.984 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:51:22 -0400 (0:00:00.047) 0:02:34.032 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:51:22 -0400 (0:00:00.037) 0:02:34.069 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext3", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "9G", "type": "lvm", "uuid": "4a67b7e5-56e4-4a75-9eda-5719be04ec64" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "1Y1RS3-4KcE-61ng-zlkk-uC1e-VKgr-w1g5I3" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:51:22 -0400 (0:00:00.323) 0:02:34.393 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002636", "end": "2024-08-17 19:51:23.134600", "rc": 0, "start": "2024-08-17 19:51:23.131964" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:51:23 -0400 (0:00:00.317) 0:02:34.711 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002606", "end": "2024-08-17 19:51:23.448883", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:51:23.446277" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:51:23 -0400 (0:00:00.308) 0:02:35.019 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:51:23 -0400 (0:00:00.078) 0:02:35.098 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:51:23 -0400 (0:00:00.035) 0:02:35.134 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.018552", "end": "2024-08-17 19:51:23.892680", "rc": 0, "start": "2024-08-17 19:51:23.874128" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:51:23 -0400 (0:00:00.329) 0:02:35.463 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.048) 0:02:35.511 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.080) 0:02:35.591 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.046) 0:02:35.637 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.317) 0:02:35.955 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.042) 0:02:35.997 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.043) 0:02:36.041 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.044) 0:02:36.085 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.041) 0:02:36.127 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.043) 0:02:36.171 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.035) 0:02:36.206 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.050) 0:02:36.257 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:51:24 -0400 (0:00:00.233) 0:02:36.491 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.091) 0:02:36.583 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.072) 0:02:36.655 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:36.691 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.045) 0:02:36.736 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.040) 0:02:36.777 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.037) 0:02:36.814 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:36.851 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.035) 0:02:36.887 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.035) 0:02:36.923 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.034) 0:02:36.957 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.037) 0:02:36.995 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:37.031 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:37.068 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.075) 0:02:37.143 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.077) 0:02:37.221 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.035) 0:02:37.257 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:37.293 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.035) 0:02:37.329 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:37.365 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.038) 0:02:37.404 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:37.440 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:51:25 -0400 (0:00:00.036) 0:02:37.476 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.077) 0:02:37.553 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.080) 0:02:37.634 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.035) 0:02:37.670 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.035) 0:02:37.705 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.034) 0:02:37.740 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.051) 0:02:37.791 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.081) 0:02:37.872 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.041) 0:02:37.913 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.040) 0:02:37.954 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.070) 0:02:38.024 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.042) 0:02:38.066 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.042) 0:02:38.109 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.034) 0:02:38.144 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.036) 0:02:38.180 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.034) 0:02:38.215 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.037) 0:02:38.253 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.035) 0:02:38.288 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.083) 0:02:38.372 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.077) 0:02:38.449 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:51:26 -0400 (0:00:00.035) 0:02:38.484 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.035) 0:02:38.520 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.037) 0:02:38.557 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.035) 0:02:38.593 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.035) 0:02:38.628 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.034) 0:02:38.662 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.033) 0:02:38.696 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.153) 0:02:38.849 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.035) 0:02:38.885 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.035) 0:02:38.920 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.035) 0:02:38.956 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.036) 0:02:38.993 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.038) 0:02:39.031 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.036) 0:02:39.068 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.039) 0:02:39.107 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.081) 0:02:39.188 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.044) 0:02:39.233 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.178) 0:02:39.411 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.041) 0:02:39.453 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:51:27 -0400 (0:00:00.044) 0:02:39.497 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.036) 0:02:39.534 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.042) 0:02:39.576 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.035) 0:02:39.612 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.038) 0:02:39.651 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.035) 0:02:39.686 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.034) 0:02:39.721 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.035) 0:02:39.756 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.035) 0:02:39.791 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.040) 0:02:39.832 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext3 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.070) 0:02:39.903 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.043) 0:02:39.947 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.042) 0:02:39.989 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.034) 0:02:40.023 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.039) 0:02:40.063 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.035) 0:02:40.098 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.050) 0:02:40.149 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:51:28 -0400 (0:00:00.051) 0:02:40.201 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938679.0579329, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938679.0579329, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1230158, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938679.0579329, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.323) 0:02:40.524 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.046) 0:02:40.571 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.036) 0:02:40.607 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.043) 0:02:40.651 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.038) 0:02:40.689 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.034) 0:02:40.724 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.045) 0:02:40.769 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.035) 0:02:40.805 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.505) 0:02:41.310 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.035) 0:02:41.346 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.035) 0:02:41.381 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.047) 0:02:41.429 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.037) 0:02:41.466 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:51:29 -0400 (0:00:00.036) 0:02:41.503 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:41.539 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.035) 0:02:41.575 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.034) 0:02:41.610 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.044) 0:02:41.654 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.045) 0:02:41.699 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.094) 0:02:41.793 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:41.830 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:41.866 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.052) 0:02:41.919 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:41.955 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:41.992 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.040) 0:02:42.032 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:42.069 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:42.105 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.036) 0:02:42.141 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.035) 0:02:42.177 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.035) 0:02:42.212 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.039) 0:02:42.251 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:51:30 -0400 (0:00:00.037) 0:02:42.289 ******* ok: [managed_node2] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.350) 0:02:42.639 ******* ok: [managed_node2] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.348) 0:02:42.987 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "9663676416" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.048) 0:02:43.035 ******* ok: [managed_node2] => { "storage_test_expected_size": "9663676416" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.038) 0:02:43.074 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.328) 0:02:43.402 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.041) 0:02:43.444 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:51:31 -0400 (0:00:00.040) 0:02:43.485 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.040) 0:02:43.526 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.041) 0:02:43.567 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.035) 0:02:43.602 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.034) 0:02:43.637 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.037) 0:02:43.674 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.036) 0:02:43.710 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.035) 0:02:43.746 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.034) 0:02:43.780 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.034) 0:02:43.815 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.034) 0:02:43.849 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.037) 0:02:43.887 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.035) 0:02:43.923 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.045) 0:02:43.969 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.039) 0:02:44.009 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.034) 0:02:44.044 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.034) 0:02:44.079 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.037) 0:02:44.117 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 9663676416, "changed": false, "failed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.040) 0:02:44.157 ******* ok: [managed_node2] => { "storage_test_expected_size": "9663676416" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.039) 0:02:44.197 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:51:32 -0400 (0:00:00.047) 0:02:44.245 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.015773", "end": "2024-08-17 19:51:33.004718", "rc": 0, "start": "2024-08-17 19:51:32.988945" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.331) 0:02:44.576 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.044) 0:02:44.621 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.046) 0:02:44.667 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.039) 0:02:44.706 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.041) 0:02:44.747 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.038) 0:02:44.786 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.038) 0:02:44.825 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.036) 0:02:44.861 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.031) 0:02:44.893 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 5g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:214 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.036) 0:02:44.930 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.115) 0:02:45.045 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.057) 0:02:45.102 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.099) 0:02:45.202 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.085) 0:02:45.288 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.036) 0:02:45.324 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.035) 0:02:45.359 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.035) 0:02:45.395 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:51:33 -0400 (0:00:00.038) 0:02:45.434 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.081) 0:02:45.516 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.036) 0:02:45.552 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext3", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.044) 0:02:45.596 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.041) 0:02:45.638 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.036) 0:02:45.675 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.036) 0:02:45.711 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.035) 0:02:45.747 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.038) 0:02:45.785 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.051) 0:02:45.837 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:51:34 -0400 (0:00:00.035) 0:02:45.873 ******* changed: [managed_node2] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:51:39 -0400 (0:00:05.106) 0:02:50.979 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:51:39 -0400 (0:00:00.036) 0:02:51.015 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938660.20387, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "1ff1203632e7c5aaed867a74eb25885b038df3b8", "ctime": 1723938660.20187, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938660.20187, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:51:39 -0400 (0:00:00.323) 0:02:51.339 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.315) 0:02:51.654 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.032) 0:02:51.687 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.052) 0:02:51.740 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.046) 0:02:51.786 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.040) 0:02:51.827 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.035) 0:02:51.862 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:51:40 -0400 (0:00:00.448) 0:02:52.311 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext3', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:51:41 -0400 (0:00:00.348) 0:02:52.659 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext3', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:51:41 -0400 (0:00:00.045) 0:02:52.704 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:51:41 -0400 (0:00:00.450) 0:02:53.155 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:51:41 -0400 (0:00:00.328) 0:02:53.483 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:51:42 -0400 (0:00:00.033) 0:02:53.517 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:228 Saturday 17 August 2024 19:51:42 -0400 (0:00:00.675) 0:02:54.192 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:51:42 -0400 (0:00:00.096) 0:02:54.289 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:51:42 -0400 (0:00:00.047) 0:02:54.336 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:51:42 -0400 (0:00:00.035) 0:02:54.372 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext3", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "4a67b7e5-56e4-4a75-9eda-5719be04ec64" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "1Y1RS3-4KcE-61ng-zlkk-uC1e-VKgr-w1g5I3" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:51:43 -0400 (0:00:00.375) 0:02:54.748 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002886", "end": "2024-08-17 19:51:43.503284", "rc": 0, "start": "2024-08-17 19:51:43.500398" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:51:43 -0400 (0:00:00.330) 0:02:55.079 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002987", "end": "2024-08-17 19:51:43.820845", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:51:43.817858" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:51:43 -0400 (0:00:00.320) 0:02:55.399 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:51:43 -0400 (0:00:00.080) 0:02:55.480 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.036) 0:02:55.516 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.017206", "end": "2024-08-17 19:51:44.291734", "rc": 0, "start": "2024-08-17 19:51:44.274528" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.352) 0:02:55.868 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.048) 0:02:55.917 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.080) 0:02:55.997 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.049) 0:02:56.046 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.339) 0:02:56.386 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.043) 0:02:56.430 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:51:44 -0400 (0:00:00.044) 0:02:56.474 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.046) 0:02:56.520 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.042) 0:02:56.562 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.041) 0:02:56.603 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.035) 0:02:56.638 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.057) 0:02:56.696 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.236) 0:02:56.932 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.040) 0:02:56.973 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.073) 0:02:57.047 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.036) 0:02:57.083 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.036) 0:02:57.119 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.035) 0:02:57.155 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.035) 0:02:57.191 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.038) 0:02:57.229 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.034) 0:02:57.264 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.040) 0:02:57.304 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.043) 0:02:57.347 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.035) 0:02:57.383 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.035) 0:02:57.419 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:51:45 -0400 (0:00:00.038) 0:02:57.457 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.074) 0:02:57.531 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.078) 0:02:57.610 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.035) 0:02:57.645 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.035) 0:02:57.680 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.035) 0:02:57.716 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.036) 0:02:57.752 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.035) 0:02:57.788 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.038) 0:02:57.826 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.090) 0:02:57.917 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.077) 0:02:57.994 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.077) 0:02:58.072 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.036) 0:02:58.108 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.035) 0:02:58.144 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.035) 0:02:58.179 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.036) 0:02:58.216 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.083) 0:02:58.300 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.051) 0:02:58.351 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.045) 0:02:58.396 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.068) 0:02:58.464 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:51:46 -0400 (0:00:00.042) 0:02:58.506 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.042) 0:02:58.548 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.034) 0:02:58.583 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.035) 0:02:58.618 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.035) 0:02:58.654 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.036) 0:02:58.690 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.038) 0:02:58.728 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.083) 0:02:58.812 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.078) 0:02:58.891 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.036) 0:02:58.927 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.036) 0:02:58.963 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.035) 0:02:58.999 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.035) 0:02:59.034 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.035) 0:02:59.069 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.038) 0:02:59.107 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.036) 0:02:59.144 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.089) 0:02:59.233 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.038) 0:02:59.271 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.036) 0:02:59.307 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.036) 0:02:59.343 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.043) 0:02:59.387 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.039) 0:02:59.427 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.035) 0:02:59.463 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:51:47 -0400 (0:00:00.038) 0:02:59.501 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.070) 0:02:59.571 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.043) 0:02:59.615 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.175) 0:02:59.790 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.042) 0:02:59.833 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.044) 0:02:59.877 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.036) 0:02:59.914 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.043) 0:02:59.957 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.035) 0:02:59.993 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.035) 0:03:00.029 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.097) 0:03:00.127 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.036) 0:03:00.164 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.035) 0:03:00.199 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.035) 0:03:00.235 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.036) 0:03:00.271 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext3 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.062) 0:03:00.333 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.046) 0:03:00.379 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.055) 0:03:00.434 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:51:48 -0400 (0:00:00.039) 0:03:00.474 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.041) 0:03:00.515 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.035) 0:03:00.551 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.047) 0:03:00.599 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.054) 0:03:00.654 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938699.3520005, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938699.3520005, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1230158, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938699.3520005, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.315) 0:03:00.970 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.047) 0:03:01.017 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.036) 0:03:01.053 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.042) 0:03:01.096 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.038) 0:03:01.135 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.034) 0:03:01.169 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.042) 0:03:01.212 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:51:49 -0400 (0:00:00.039) 0:03:01.251 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.498) 0:03:01.749 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.035) 0:03:01.785 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.035) 0:03:01.821 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.048) 0:03:01.869 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.035) 0:03:01.905 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.038) 0:03:01.944 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.036) 0:03:01.980 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.035) 0:03:02.016 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.037) 0:03:02.053 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.045) 0:03:02.099 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.042) 0:03:02.141 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.038) 0:03:02.179 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.035) 0:03:02.215 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.034) 0:03:02.249 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.036) 0:03:02.286 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.036) 0:03:02.322 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.036) 0:03:02.359 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.039) 0:03:02.398 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.036) 0:03:02.435 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:51:50 -0400 (0:00:00.041) 0:03:02.477 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.042) 0:03:02.519 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.034) 0:03:02.554 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.035) 0:03:02.590 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.038) 0:03:02.628 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.037) 0:03:02.665 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.322) 0:03:02.988 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.324) 0:03:03.313 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.047) 0:03:03.360 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:51:51 -0400 (0:00:00.041) 0:03:03.401 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.328) 0:03:03.729 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.040) 0:03:03.770 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.041) 0:03:03.812 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.041) 0:03:03.853 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.041) 0:03:03.894 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.035) 0:03:03.930 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.035) 0:03:03.965 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.033) 0:03:03.999 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.037) 0:03:04.036 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.035) 0:03:04.071 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.034) 0:03:04.106 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.035) 0:03:04.141 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.034) 0:03:04.176 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.034) 0:03:04.211 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.091) 0:03:04.302 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.036) 0:03:04.339 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.036) 0:03:04.375 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.036) 0:03:04.411 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.034) 0:03:04.446 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:51:52 -0400 (0:00:00.035) 0:03:04.482 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.042) 0:03:04.524 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.050) 0:03:04.575 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.049) 0:03:04.625 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.017206", "end": "2024-08-17 19:51:53.384678", "rc": 0, "start": "2024-08-17 19:51:53.367472" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.333) 0:03:04.959 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.044) 0:03:05.004 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.046) 0:03:05.050 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.038) 0:03:05.089 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.038) 0:03:05.128 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.038) 0:03:05.167 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.041) 0:03:05.208 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.036) 0:03:05.245 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.033) 0:03:05.279 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:231 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.035) 0:03:05.314 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.110) 0:03:05.425 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:51:53 -0400 (0:00:00.054) 0:03:05.480 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.045) 0:03:05.525 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.101) 0:03:05.626 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.036) 0:03:05.663 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.034) 0:03:05.698 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.035) 0:03:05.733 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.035) 0:03:05.769 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.084) 0:03:05.853 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.036) 0:03:05.890 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.044) 0:03:05.935 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.041) 0:03:05.976 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.036) 0:03:06.013 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.036) 0:03:06.050 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.036) 0:03:06.086 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.036) 0:03:06.123 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.052) 0:03:06.175 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:51:54 -0400 (0:00:00.035) 0:03:06.210 ******* changed: [managed_node2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "fstype": "ext3", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:51:59 -0400 (0:00:04.515) 0:03:10.726 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:51:59 -0400 (0:00:00.035) 0:03:10.762 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938660.20387, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "1ff1203632e7c5aaed867a74eb25885b038df3b8", "ctime": 1723938660.20187, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938660.20187, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:51:59 -0400 (0:00:00.323) 0:03:11.085 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:51:59 -0400 (0:00:00.319) 0:03:11.405 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:51:59 -0400 (0:00:00.032) 0:03:11.438 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "fstype": "ext3", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:51:59 -0400 (0:00:00.046) 0:03:11.484 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:52:00 -0400 (0:00:00.045) 0:03:11.529 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:52:00 -0400 (0:00:00.104) 0:03:11.634 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'state': u'absent', u'fstype': u'ext3', u'path': u'/opt/test1'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "fstype": "ext3", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:52:00 -0400 (0:00:00.344) 0:03:11.978 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:52:00 -0400 (0:00:00.449) 0:03:12.427 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:52:00 -0400 (0:00:00.037) 0:03:12.465 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:52:00 -0400 (0:00:00.036) 0:03:12.501 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:52:01 -0400 (0:00:00.464) 0:03:12.966 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:52:01 -0400 (0:00:00.368) 0:03:13.334 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:52:01 -0400 (0:00:00.034) 0:03:13.369 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:244 Saturday 17 August 2024 19:52:02 -0400 (0:00:00.705) 0:03:14.075 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:52:02 -0400 (0:00:00.084) 0:03:14.159 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:52:02 -0400 (0:00:00.047) 0:03:14.207 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:52:02 -0400 (0:00:00.037) 0:03:14.245 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.333) 0:03:14.578 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002865", "end": "2024-08-17 19:52:03.342323", "rc": 0, "start": "2024-08-17 19:52:03.339458" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.340) 0:03:14.919 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002781", "end": "2024-08-17 19:52:03.675345", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:52:03.672564" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.330) 0:03:15.249 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.076) 0:03:15.326 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.036) 0:03:15.362 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.038) 0:03:15.401 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:52:03 -0400 (0:00:00.035) 0:03:15.436 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.079) 0:03:15.516 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.046) 0:03:15.562 ******* TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.034) 0:03:15.597 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.041) 0:03:15.638 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.044) 0:03:15.682 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.045) 0:03:15.728 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.042) 0:03:15.770 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.045) 0:03:15.815 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.049) 0:03:15.865 ******* TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.032) 0:03:15.897 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.267) 0:03:16.164 ******* TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.033) 0:03:16.197 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.072) 0:03:16.270 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.035) 0:03:16.305 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.035) 0:03:16.340 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.094) 0:03:16.434 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.036) 0:03:16.470 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:52:04 -0400 (0:00:00.035) 0:03:16.506 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.036) 0:03:16.542 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.035) 0:03:16.578 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.035) 0:03:16.614 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.038) 0:03:16.652 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.035) 0:03:16.688 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.036) 0:03:16.724 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.074) 0:03:16.799 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.090) 0:03:16.889 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.035) 0:03:16.925 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.035) 0:03:16.960 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:16.994 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.029 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.037) 0:03:17.067 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.102 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.136 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.074) 0:03:17.210 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.074) 0:03:17.285 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.319 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.354 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.389 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.034) 0:03:17.424 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:52:05 -0400 (0:00:00.081) 0:03:17.505 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.041) 0:03:17.547 ******* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.031) 0:03:17.579 ******* TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.031) 0:03:17.611 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.037) 0:03:17.649 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.081) 0:03:17.730 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.076) 0:03:17.806 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.038) 0:03:17.845 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.038) 0:03:17.883 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.048) 0:03:17.931 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.035) 0:03:17.966 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.035) 0:03:18.002 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.035) 0:03:18.037 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.038) 0:03:18.075 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.086) 0:03:18.162 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.035) 0:03:18.198 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.034) 0:03:18.232 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.036) 0:03:18.269 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.034) 0:03:18.304 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.035) 0:03:18.339 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.034) 0:03:18.373 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.034) 0:03:18.408 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:52:06 -0400 (0:00:00.069) 0:03:18.477 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.042) 0:03:18.520 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.232) 0:03:18.752 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.041) 0:03:18.794 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.044) 0:03:18.838 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.038) 0:03:18.877 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.029) 0:03:18.907 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.047) 0:03:18.954 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.034) 0:03:18.989 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.034) 0:03:19.023 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.035) 0:03:19.058 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.037) 0:03:19.096 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.035) 0:03:19.132 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.035) 0:03:19.167 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.060) 0:03:19.228 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.034) 0:03:19.262 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.042) 0:03:19.304 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.037) 0:03:19.342 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.040) 0:03:19.382 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.035) 0:03:19.418 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.038) 0:03:19.456 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:52:07 -0400 (0:00:00.037) 0:03:19.494 ******* ok: [managed_node2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.308) 0:03:19.803 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.035) 0:03:19.838 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.045) 0:03:19.884 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.032) 0:03:19.916 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.043) 0:03:19.959 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.044) 0:03:20.004 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.029) 0:03:20.034 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:52:08 -0400 (0:00:00.034) 0:03:20.069 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.500) 0:03:20.569 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.036) 0:03:20.606 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:20.641 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.029) 0:03:20.671 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.035) 0:03:20.706 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:20.741 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.030) 0:03:20.771 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.033) 0:03:20.805 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.029) 0:03:20.834 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.044) 0:03:20.879 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.041) 0:03:20.920 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.033) 0:03:20.954 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:20.988 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.050) 0:03:21.038 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.036) 0:03:21.075 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:21.109 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:21.144 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.033) 0:03:21.178 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:21.213 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.037) 0:03:21.251 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.035) 0:03:21.286 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.035) 0:03:21.322 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:21.357 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.034) 0:03:21.392 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.035) 0:03:21.428 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:52:09 -0400 (0:00:00.037) 0:03:21.466 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.090) 0:03:21.556 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.592 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.038) 0:03:21.630 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:21.665 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.700 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.736 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.037) 0:03:21.773 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.808 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.844 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.879 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:21.915 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:21.950 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.038) 0:03:21.988 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.038) 0:03:22.026 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.044) 0:03:22.071 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:22.106 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:22.140 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:22.175 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.037) 0:03:22.212 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.037) 0:03:22.249 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:22.284 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:22.320 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.034) 0:03:22.354 ******* ok: [managed_node2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.039) 0:03:22.393 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.042) 0:03:22.436 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:22.471 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:52:10 -0400 (0:00:00.035) 0:03:22.506 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.035) 0:03:22.542 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.035) 0:03:22.578 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.035) 0:03:22.613 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.037) 0:03:22.651 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.035) 0:03:22.687 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.035) 0:03:22.722 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.036) 0:03:22.759 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.031) 0:03:22.790 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a LVM logical volume with for ext2 FS size 5g] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:249 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.036) 0:03:22.827 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.123) 0:03:22.951 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.056) 0:03:23.008 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.041) 0:03:23.050 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.097) 0:03:23.147 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.035) 0:03:23.182 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.037) 0:03:23.219 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.034) 0:03:23.254 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.037) 0:03:23.291 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.079) 0:03:23.371 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.037) 0:03:23.408 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext2", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.043) 0:03:23.452 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:52:11 -0400 (0:00:00.037) 0:03:23.490 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:52:12 -0400 (0:00:00.034) 0:03:23.525 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:52:12 -0400 (0:00:00.035) 0:03:23.560 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:52:12 -0400 (0:00:00.035) 0:03:23.595 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:52:12 -0400 (0:00:00.037) 0:03:23.633 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:52:12 -0400 (0:00:00.051) 0:03:23.685 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:52:12 -0400 (0:00:00.032) 0:03:23.717 ******* changed: [managed_node2] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:52:16 -0400 (0:00:04.662) 0:03:28.380 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:52:16 -0400 (0:00:00.035) 0:03:28.415 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938720.809072, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "72884e3f126482c2d28276ff7c57744fa95eff91", "ctime": 1723938720.3970706, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938720.3970706, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1229, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.322) 0:03:28.737 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.329) 0:03:29.067 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.033) 0:03:29.101 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.049) 0:03:29.150 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.045) 0:03:29.196 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.051) 0:03:29.247 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:52:17 -0400 (0:00:00.034) 0:03:29.282 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:52:18 -0400 (0:00:00.463) 0:03:29.745 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext2', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:52:18 -0400 (0:00:00.349) 0:03:30.095 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext2', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:52:18 -0400 (0:00:00.045) 0:03:30.141 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:52:19 -0400 (0:00:00.453) 0:03:30.595 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:52:19 -0400 (0:00:00.328) 0:03:30.923 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:52:19 -0400 (0:00:00.033) 0:03:30.957 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:265 Saturday 17 August 2024 19:52:20 -0400 (0:00:00.675) 0:03:31.633 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:52:20 -0400 (0:00:00.092) 0:03:31.725 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:52:20 -0400 (0:00:00.048) 0:03:31.774 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:52:20 -0400 (0:00:00.034) 0:03:31.808 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext2", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "cf46f49b-2762-463c-bd7d-d37dde1405fc" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "o3vJdQ-6JKX-k1VQ-9fYJ-igHX-p2GR-Kx03pg" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:52:20 -0400 (0:00:00.325) 0:03:32.134 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002764", "end": "2024-08-17 19:52:20.877976", "rc": 0, "start": "2024-08-17 19:52:20.875212" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:52:20 -0400 (0:00:00.317) 0:03:32.452 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002671", "end": "2024-08-17 19:52:21.198803", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:52:21.196132" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.322) 0:03:32.775 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.080) 0:03:32.856 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.036) 0:03:32.893 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.017097", "end": "2024-08-17 19:52:21.655771", "rc": 0, "start": "2024-08-17 19:52:21.638674" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.337) 0:03:33.230 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.049) 0:03:33.280 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.093) 0:03:33.374 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:52:21 -0400 (0:00:00.049) 0:03:33.423 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.314) 0:03:33.738 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.043) 0:03:33.781 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.101) 0:03:33.883 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.046) 0:03:33.930 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.043) 0:03:33.974 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.043) 0:03:34.018 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.036) 0:03:34.054 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.052) 0:03:34.106 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.238) 0:03:34.345 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.055) 0:03:34.400 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:52:22 -0400 (0:00:00.075) 0:03:34.476 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.037) 0:03:34.513 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:34.549 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:34.586 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:34.622 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.039) 0:03:34.661 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.037) 0:03:34.698 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.035) 0:03:34.734 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.035) 0:03:34.770 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:34.806 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:34.842 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.039) 0:03:34.882 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.076) 0:03:34.958 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.077) 0:03:35.036 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.038) 0:03:35.074 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:35.111 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:35.147 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.034) 0:03:35.182 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.035) 0:03:35.217 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.035) 0:03:35.252 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.039) 0:03:35.291 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.077) 0:03:35.369 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.091) 0:03:35.460 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:52:23 -0400 (0:00:00.036) 0:03:35.497 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.035) 0:03:35.532 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.035) 0:03:35.567 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.035) 0:03:35.603 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.084) 0:03:35.687 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.042) 0:03:35.730 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.041) 0:03:35.771 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.070) 0:03:35.842 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.041) 0:03:35.884 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.043) 0:03:35.928 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.035) 0:03:35.963 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.036) 0:03:35.999 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.034) 0:03:36.034 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.039) 0:03:36.074 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.035) 0:03:36.110 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.080) 0:03:36.190 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.135) 0:03:36.326 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.037) 0:03:36.363 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.036) 0:03:36.399 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.040) 0:03:36.439 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:52:24 -0400 (0:00:00.044) 0:03:36.484 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:36.521 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.039) 0:03:36.560 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:36.597 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.088) 0:03:36.685 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.039) 0:03:36.725 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:36.761 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:36.798 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:36.834 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:36.870 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.037) 0:03:36.908 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.040) 0:03:36.948 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.072) 0:03:37.020 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.043) 0:03:37.064 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.179) 0:03:37.244 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.043) 0:03:37.287 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.044) 0:03:37.332 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:37.369 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.043) 0:03:37.413 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.036) 0:03:37.449 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:52:25 -0400 (0:00:00.047) 0:03:37.496 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.039) 0:03:37.536 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.036) 0:03:37.572 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.035) 0:03:37.608 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.035) 0:03:37.643 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.036) 0:03:37.680 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext2 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.064) 0:03:37.745 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.044) 0:03:37.789 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.043) 0:03:37.833 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.036) 0:03:37.870 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.041) 0:03:37.911 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.036) 0:03:37.948 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.051) 0:03:37.999 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.053) 0:03:38.052 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938736.7651253, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938736.7651253, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1249773, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938736.7651253, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.338) 0:03:38.391 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.046) 0:03:38.438 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:52:26 -0400 (0:00:00.037) 0:03:38.475 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.055) 0:03:38.530 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.041) 0:03:38.572 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.036) 0:03:38.608 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.044) 0:03:38.652 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.036) 0:03:38.689 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.537) 0:03:39.226 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.036) 0:03:39.263 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.036) 0:03:39.300 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.049) 0:03:39.349 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.037) 0:03:39.387 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.039) 0:03:39.427 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.036) 0:03:39.464 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:52:27 -0400 (0:00:00.036) 0:03:39.500 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.037) 0:03:39.538 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.054) 0:03:39.592 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.042) 0:03:39.635 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.040) 0:03:39.675 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:39.712 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:39.748 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.037) 0:03:39.785 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:39.822 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.037) 0:03:39.859 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.039) 0:03:39.899 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:39.936 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:39.973 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:40.010 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.037) 0:03:40.047 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.036) 0:03:40.084 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.039) 0:03:40.123 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.037) 0:03:40.161 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:52:28 -0400 (0:00:00.314) 0:03:40.475 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.332) 0:03:40.808 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.048) 0:03:40.857 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.040) 0:03:40.897 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.314) 0:03:41.212 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.041) 0:03:41.254 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.042) 0:03:41.297 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.041) 0:03:41.338 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.042) 0:03:41.381 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.036) 0:03:41.418 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.035) 0:03:41.453 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:52:29 -0400 (0:00:00.035) 0:03:41.489 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.039) 0:03:41.529 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.037) 0:03:41.566 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.038) 0:03:41.605 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.045) 0:03:41.651 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.036) 0:03:41.687 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.037) 0:03:41.724 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.039) 0:03:41.763 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.036) 0:03:41.800 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.036) 0:03:41.837 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.037) 0:03:41.874 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.036) 0:03:41.911 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.035) 0:03:41.947 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.044) 0:03:41.991 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.041) 0:03:42.033 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.049) 0:03:42.082 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.016496", "end": "2024-08-17 19:52:30.847608", "rc": 0, "start": "2024-08-17 19:52:30.831112" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.336) 0:03:42.419 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:52:30 -0400 (0:00:00.045) 0:03:42.465 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.047) 0:03:42.512 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.039) 0:03:42.551 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.039) 0:03:42.591 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.042) 0:03:42.633 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.053) 0:03:42.687 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.036) 0:03:42.723 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.033) 0:03:42.757 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 9g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:268 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.037) 0:03:42.794 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.193) 0:03:42.988 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.058) 0:03:43.046 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.044) 0:03:43.090 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.090) 0:03:43.181 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.036) 0:03:43.217 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.035) 0:03:43.252 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.036) 0:03:43.289 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.038) 0:03:43.327 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.081) 0:03:43.409 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.035) 0:03:43.445 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext2", "mount_point": "/opt/test1", "name": "test1", "size": "9g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:52:31 -0400 (0:00:00.045) 0:03:43.490 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.039) 0:03:43.530 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.036) 0:03:43.567 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.036) 0:03:43.603 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.036) 0:03:43.639 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.036) 0:03:43.675 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.070) 0:03:43.746 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:52:32 -0400 (0:00:00.033) 0:03:43.779 ******* changed: [managed_node2] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:52:37 -0400 (0:00:05.241) 0:03:49.021 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:52:37 -0400 (0:00:00.036) 0:03:49.057 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938738.508131, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "c6abf0a388bb293ed38f7a11686eb491e2373da7", "ctime": 1723938738.5061312, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938738.5061312, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:52:37 -0400 (0:00:00.323) 0:03:49.381 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.324) 0:03:49.705 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.033) 0:03:49.738 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.048) 0:03:49.787 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.046) 0:03:49.833 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.052) 0:03:49.886 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.038) 0:03:49.924 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:52:38 -0400 (0:00:00.448) 0:03:50.373 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext2', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:52:39 -0400 (0:00:00.340) 0:03:50.713 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext2', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:52:39 -0400 (0:00:00.047) 0:03:50.761 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:52:39 -0400 (0:00:00.439) 0:03:51.200 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:52:40 -0400 (0:00:00.313) 0:03:51.514 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:52:40 -0400 (0:00:00.034) 0:03:51.549 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:282 Saturday 17 August 2024 19:52:41 -0400 (0:00:01.766) 0:03:53.315 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:52:41 -0400 (0:00:00.155) 0:03:53.470 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:52:42 -0400 (0:00:00.049) 0:03:53.520 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:52:42 -0400 (0:00:00.037) 0:03:53.558 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext2", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "9G", "type": "lvm", "uuid": "cf46f49b-2762-463c-bd7d-d37dde1405fc" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "o3vJdQ-6JKX-k1VQ-9fYJ-igHX-p2GR-Kx03pg" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:52:42 -0400 (0:00:00.324) 0:03:53.882 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002708", "end": "2024-08-17 19:52:42.637494", "rc": 0, "start": "2024-08-17 19:52:42.634786" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:52:42 -0400 (0:00:00.330) 0:03:54.212 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002622", "end": "2024-08-17 19:52:42.954552", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:52:42.951930" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.312) 0:03:54.524 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.078) 0:03:54.603 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.036) 0:03:54.640 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.020275", "end": "2024-08-17 19:52:43.424661", "rc": 0, "start": "2024-08-17 19:52:43.404386" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.369) 0:03:55.009 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.048) 0:03:55.057 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.081) 0:03:55.138 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:52:43 -0400 (0:00:00.046) 0:03:55.185 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.333) 0:03:55.518 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.042) 0:03:55.560 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.043) 0:03:55.603 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.044) 0:03:55.648 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.043) 0:03:55.692 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.043) 0:03:55.735 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.036) 0:03:55.772 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.052) 0:03:55.825 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.250) 0:03:56.075 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.041) 0:03:56.116 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.072) 0:03:56.189 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.035) 0:03:56.224 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.038) 0:03:56.263 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.036) 0:03:56.300 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.035) 0:03:56.335 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.035) 0:03:56.370 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.035) 0:03:56.406 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.035) 0:03:56.442 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:52:44 -0400 (0:00:00.038) 0:03:56.480 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.035) 0:03:56.515 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.035) 0:03:56.551 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.036) 0:03:56.587 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.075) 0:03:56.663 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.075) 0:03:56.738 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.035) 0:03:56.774 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.034) 0:03:56.809 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.037) 0:03:56.847 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.093) 0:03:56.941 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.036) 0:03:56.977 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.036) 0:03:57.014 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.049) 0:03:57.063 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.078) 0:03:57.141 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.077) 0:03:57.218 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.035) 0:03:57.254 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.034) 0:03:57.288 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.037) 0:03:57.326 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.036) 0:03:57.362 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.080) 0:03:57.442 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:52:45 -0400 (0:00:00.042) 0:03:57.485 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.043) 0:03:57.528 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.069) 0:03:57.597 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.041) 0:03:57.639 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.044) 0:03:57.683 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:57.719 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.037) 0:03:57.756 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:57.792 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.036) 0:03:57.828 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.036) 0:03:57.865 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.083) 0:03:57.948 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.078) 0:03:58.027 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.043) 0:03:58.071 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.040) 0:03:58.111 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.037) 0:03:58.149 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.036) 0:03:58.185 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:58.220 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.034) 0:03:58.255 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:58.291 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.090) 0:03:58.381 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:58.416 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:58.452 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:52:46 -0400 (0:00:00.035) 0:03:58.487 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.038) 0:03:58.526 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.035) 0:03:58.561 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.035) 0:03:58.597 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.035) 0:03:58.633 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.070) 0:03:58.704 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.044) 0:03:58.748 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.174) 0:03:58.923 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.040) 0:03:58.964 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.043) 0:03:59.008 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.035) 0:03:59.044 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.123) 0:03:59.168 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.036) 0:03:59.204 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.036) 0:03:59.241 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.036) 0:03:59.277 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.035) 0:03:59.313 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.035) 0:03:59.348 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.039) 0:03:59.387 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.037) 0:03:59.425 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext2 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:52:47 -0400 (0:00:00.062) 0:03:59.488 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.043) 0:03:59.532 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.043) 0:03:59.575 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.035) 0:03:59.611 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.044) 0:03:59.656 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.037) 0:03:59.693 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.049) 0:03:59.743 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.051) 0:03:59.795 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938757.4091942, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938757.4091942, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1249773, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938757.4091942, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.320) 0:04:00.115 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.061) 0:04:00.176 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.036) 0:04:00.213 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.043) 0:04:00.256 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.041) 0:04:00.298 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.036) 0:04:00.335 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.042) 0:04:00.378 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:52:48 -0400 (0:00:00.035) 0:04:00.413 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.495) 0:04:00.909 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.035) 0:04:00.945 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.036) 0:04:00.981 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.049) 0:04:01.031 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.035) 0:04:01.066 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.035) 0:04:01.102 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.035) 0:04:01.137 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.041) 0:04:01.178 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.044) 0:04:01.223 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.044) 0:04:01.268 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.042) 0:04:01.310 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.036) 0:04:01.347 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.035) 0:04:01.382 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.035) 0:04:01.417 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.040) 0:04:01.457 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:52:49 -0400 (0:00:00.036) 0:04:01.494 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.036) 0:04:01.530 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.036) 0:04:01.567 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.035) 0:04:01.603 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.035) 0:04:01.638 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.040) 0:04:01.679 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.036) 0:04:01.716 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.036) 0:04:01.752 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.036) 0:04:01.789 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.036) 0:04:01.825 ******* ok: [managed_node2] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.319) 0:04:02.145 ******* ok: [managed_node2] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:52:50 -0400 (0:00:00.326) 0:04:02.471 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "9663676416" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.049) 0:04:02.520 ******* ok: [managed_node2] => { "storage_test_expected_size": "9663676416" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.040) 0:04:02.560 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.325) 0:04:02.886 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.041) 0:04:02.927 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.039) 0:04:02.967 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.040) 0:04:03.008 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.040) 0:04:03.049 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.037) 0:04:03.086 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.035) 0:04:03.122 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.035) 0:04:03.158 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.034) 0:04:03.193 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.035) 0:04:03.228 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.046) 0:04:03.275 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.037) 0:04:03.312 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.036) 0:04:03.349 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.035) 0:04:03.384 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.034) 0:04:03.419 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.035) 0:04:03.454 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:52:51 -0400 (0:00:00.034) 0:04:03.489 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.096) 0:04:03.585 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.036) 0:04:03.622 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.036) 0:04:03.658 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 9663676416, "changed": false, "failed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.040) 0:04:03.699 ******* ok: [managed_node2] => { "storage_test_expected_size": "9663676416" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.039) 0:04:03.739 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.049) 0:04:03.788 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.019017", "end": "2024-08-17 19:52:52.554642", "rc": 0, "start": "2024-08-17 19:52:52.535625" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.337) 0:04:04.125 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.045) 0:04:04.171 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.047) 0:04:04.218 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.040) 0:04:04.259 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.052) 0:04:04.312 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.039) 0:04:04.352 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.039) 0:04:04.391 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.035) 0:04:04.427 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.032) 0:04:04.459 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 5g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:285 Saturday 17 August 2024 19:52:52 -0400 (0:00:00.039) 0:04:04.499 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.139) 0:04:04.638 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.054) 0:04:04.693 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.042) 0:04:04.735 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.091) 0:04:04.826 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.035) 0:04:04.862 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.035) 0:04:04.897 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.035) 0:04:04.933 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.035) 0:04:04.968 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.081) 0:04:05.050 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.035) 0:04:05.086 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext2", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.043) 0:04:05.129 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.039) 0:04:05.169 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.035) 0:04:05.204 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.037) 0:04:05.242 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.038) 0:04:05.281 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.046) 0:04:05.327 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.054) 0:04:05.382 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:52:53 -0400 (0:00:00.032) 0:04:05.415 ******* changed: [managed_node2] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:52:59 -0400 (0:00:05.202) 0:04:10.617 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.035) 0:04:10.653 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938738.508131, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "c6abf0a388bb293ed38f7a11686eb491e2373da7", "ctime": 1723938738.5061312, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938738.5061312, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.325) 0:04:10.978 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.324) 0:04:11.302 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.032) 0:04:11.335 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.050) 0:04:11.385 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.045) 0:04:11.431 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:52:59 -0400 (0:00:00.049) 0:04:11.480 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:53:00 -0400 (0:00:00.037) 0:04:11.518 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:53:00 -0400 (0:00:00.458) 0:04:11.976 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext2', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:53:00 -0400 (0:00:00.345) 0:04:12.322 ******* skipping: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'group': None, u'dump': 0, u'passno': 0, u'fstype': u'ext2', u'state': u'mounted', u'mode': None, u'owner': None, u'path': u'/opt/test1', u'opts': u'defaults'}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:53:00 -0400 (0:00:00.045) 0:04:12.368 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:53:01 -0400 (0:00:00.484) 0:04:12.853 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:53:01 -0400 (0:00:00.402) 0:04:13.255 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:53:01 -0400 (0:00:00.034) 0:04:13.290 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:299 Saturday 17 August 2024 19:53:02 -0400 (0:00:00.755) 0:04:14.045 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:53:02 -0400 (0:00:00.103) 0:04:14.149 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:53:02 -0400 (0:00:00.049) 0:04:14.198 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:53:02 -0400 (0:00:00.037) 0:04:14.236 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext2", "label": "", "mountpoint": "/opt/test1", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "cf46f49b-2762-463c-bd7d-d37dde1405fc" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "o3vJdQ-6JKX-k1VQ-9fYJ-igHX-p2GR-Kx03pg" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:53:03 -0400 (0:00:00.353) 0:04:14.590 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002892", "end": "2024-08-17 19:53:03.351848", "rc": 0, "start": "2024-08-17 19:53:03.348956" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:53:03 -0400 (0:00:00.338) 0:04:14.929 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002703", "end": "2024-08-17 19:53:03.698338", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:53:03.695635" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:53:03 -0400 (0:00:00.346) 0:04:15.275 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:53:03 -0400 (0:00:00.079) 0:04:15.354 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:53:03 -0400 (0:00:00.037) 0:04:15.392 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "vgs", "--noheadings", "--binary", "-o", "shared", "foo" ], "delta": "0:00:00.017433", "end": "2024-08-17 19:53:04.170756", "rc": 0, "start": "2024-08-17 19:53:04.153323" } STDOUT: 0 TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.360) 0:04:15.752 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.049) 0:04:15.802 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.082) 0:04:15.885 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.049) 0:04:15.934 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.344) 0:04:16.278 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.042) 0:04:16.321 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.044) 0:04:16.366 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.047) 0:04:16.414 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.045) 0:04:16.459 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:53:04 -0400 (0:00:00.044) 0:04:16.504 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:16.541 ******* ok: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.052) 0:04:16.593 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.267) 0:04:16.860 ******* skipping: [managed_node2] => (item=/dev/sda) => { "ansible_loop_var": "st_pool_pv", "changed": false, "skip_reason": "Conditional result was False", "st_pool_pv": "/dev/sda" } TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.042) 0:04:16.903 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.074) 0:04:16.977 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.014 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.038) 0:04:17.052 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.089 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.125 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.162 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.198 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.235 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.038) 0:04:17.273 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.310 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.346 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.036) 0:04:17.383 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:53:05 -0400 (0:00:00.078) 0:04:17.461 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.176) 0:04:17.638 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.047) 0:04:17.686 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.035) 0:04:17.722 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.036) 0:04:17.758 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.038) 0:04:17.796 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.036) 0:04:17.833 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.036) 0:04:17.869 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.035) 0:04:17.905 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.078) 0:04:17.984 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.077) 0:04:18.062 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.035) 0:04:18.098 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.035) 0:04:18.133 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.039) 0:04:18.172 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.036) 0:04:18.209 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.081) 0:04:18.290 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.041) 0:04:18.331 ******* skipping: [managed_node2] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.044) 0:04:18.376 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml for managed_node2 TASK [Set variables used by tests] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:2 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.070) 0:04:18.446 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:6 Saturday 17 August 2024 19:53:06 -0400 (0:00:00.041) 0:04:18.488 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:14 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.043) 0:04:18.532 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:23 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.035) 0:04:18.567 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:32 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.037) 0:04:18.604 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-crypttab.yml:41 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.035) 0:04:18.640 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:18.676 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.046) 0:04:18.723 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.084) 0:04:18.807 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.078) 0:04:18.886 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:18.922 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:18.959 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.038) 0:04:18.997 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.034 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.035) 0:04:19.069 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.106 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.142 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.091) 0:04:19.233 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.270 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.306 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.035) 0:04:19.342 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.038) 0:04:19.381 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.417 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.454 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:53:07 -0400 (0:00:00.036) 0:04:19.490 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.071) 0:04:19.562 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.045) 0:04:19.608 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.255) 0:04:19.863 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.042) 0:04:19.905 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.046) 0:04:19.952 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.037) 0:04:19.989 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.047) 0:04:20.036 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.037) 0:04:20.074 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.036) 0:04:20.111 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.036) 0:04:20.148 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.036) 0:04:20.184 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.036) 0:04:20.220 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.039) 0:04:20.260 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.037) 0:04:20.297 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext2 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.064) 0:04:20.362 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.044) 0:04:20.407 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.044) 0:04:20.451 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:53:08 -0400 (0:00:00.039) 0:04:20.491 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.042) 0:04:20.533 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.037) 0:04:20.571 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.050) 0:04:20.621 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.052) 0:04:20.674 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938778.9922662, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1723938778.9922662, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1249773, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1723938778.9922662, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.339) 0:04:21.013 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.047) 0:04:21.061 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.037) 0:04:21.099 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.044) 0:04:21.143 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.042) 0:04:21.186 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.037) 0:04:21.223 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.043) 0:04:21.267 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:53:09 -0400 (0:00:00.036) 0:04:21.304 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.500) 0:04:21.804 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.041) 0:04:21.846 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.037) 0:04:21.883 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.049) 0:04:21.932 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.037) 0:04:21.970 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.036) 0:04:22.007 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.036) 0:04:22.043 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.036) 0:04:22.080 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.039) 0:04:22.119 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.045) 0:04:22.165 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.044) 0:04:22.209 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.035) 0:04:22.245 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.036) 0:04:22.282 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.036) 0:04:22.318 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.040) 0:04:22.358 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.039) 0:04:22.397 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.037) 0:04:22.435 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.037) 0:04:22.472 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:53:10 -0400 (0:00:00.037) 0:04:22.509 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.037) 0:04:22.546 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.040) 0:04:22.587 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.037) 0:04:22.624 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.036) 0:04:22.661 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.037) 0:04:22.698 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.037) 0:04:22.736 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.349) 0:04:23.085 ******* ok: [managed_node2] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.330) 0:04:23.416 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.047) 0:04:23.464 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:53:11 -0400 (0:00:00.040) 0:04:23.504 ******* ok: [managed_node2] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.326) 0:04:23.830 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.056) 0:04:23.887 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.043) 0:04:23.930 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.042) 0:04:23.973 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.042) 0:04:24.015 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.038) 0:04:24.053 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.090 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.127 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.164 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.200 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.237 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.102) 0:04:24.339 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.037) 0:04:24.376 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.413 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.037) 0:04:24.450 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:53:12 -0400 (0:00:00.036) 0:04:24.486 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.036) 0:04:24.523 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.039) 0:04:24.562 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.037) 0:04:24.600 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.037) 0:04:24.637 ******* ok: [managed_node2] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.041) 0:04:24.679 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.039) 0:04:24.719 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.050) 0:04:24.769 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.019633", "end": "2024-08-17 19:53:13.548700", "rc": 0, "start": "2024-08-17 19:53:13.529067" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.353) 0:04:25.122 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.046) 0:04:25.168 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.050) 0:04:25.218 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.041) 0:04:25.260 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.040) 0:04:25.300 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.039) 0:04:25.340 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.039) 0:04:25.379 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.038) 0:04:25.417 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.033) 0:04:25.451 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:302 Saturday 17 August 2024 19:53:13 -0400 (0:00:00.039) 0:04:25.491 ******* TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.150) 0:04:25.642 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.055) 0:04:25.697 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.044) 0:04:25.742 ******* skipping: [managed_node2] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed_node2] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed_node2] => (item=CentOS_7.yml) => { "ansible_facts": { "__storage_blivet_diskvolume_mkfs_option_map": { "ext2": "-F", "ext3": "-F", "ext4": "-F" }, "blivet_package_list": [ "python-enum34", "python-blivet3", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "{{ 'libblockdev-s390' if ansible_architecture == 's390x' else 'libblockdev' }}" ] }, "ansible_included_var_files": [ "/tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/vars/CentOS_7.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.yml" } skipping: [managed_node2] => (item=CentOS_7.9.yml) => { "ansible_loop_var": "item", "changed": false, "item": "CentOS_7.9.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if system is ostree] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:25 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.091) 0:04:25.834 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set flag to indicate system is ostree] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:30 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.035) 0:04:25.870 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Define an empty list of pools to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.035) 0:04:25.906 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Define an empty list of volumes to be used in testing] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.043) 0:04:25.949 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Include the appropriate provider tasks] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.040) 0:04:25.990 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for managed_node2 TASK [fedora.linux_system_roles.storage : Make sure blivet is available] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.085) 0:04:26.076 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Show storage_pools] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:9 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.036) 0:04:26.113 ******* ok: [managed_node2] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [fedora.linux_system_roles.storage : Show storage_volumes] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:14 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.044) 0:04:26.158 ******* ok: [managed_node2] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : Get required packages] *************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:19 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.039) 0:04:26.197 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Enable copr repositories if needed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:31 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.039) 0:04:26.236 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Make sure required packages are installed] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:37 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.038) 0:04:26.275 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Get service facts] ******************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:51 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.036) 0:04:26.311 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:57 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.037) 0:04:26.348 ******* ok: [managed_node2] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:63 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.053) 0:04:26.402 ******* TASK [fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 Saturday 17 August 2024 19:53:14 -0400 (0:00:00.033) 0:04:26.435 ******* changed: [managed_node2] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "fstype": "ext2", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:83 Saturday 17 August 2024 19:53:19 -0400 (0:00:04.476) 0:04:30.911 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Check if /etc/fstab is present] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Saturday 17 August 2024 19:53:19 -0400 (0:00:00.036) 0:04:30.948 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723938738.508131, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "c6abf0a388bb293ed38f7a11686eb491e2373da7", "ctime": 1723938738.5061312, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263550, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0644", "mtime": 1723938738.5061312, "nlink": 1, "path": "/etc/fstab", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 1280, "uid": 0, "version": "1736133796", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Add fingerprint to /etc/fstab if present] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:95 Saturday 17 August 2024 19:53:19 -0400 (0:00:00.329) 0:04:31.278 ******* ok: [managed_node2] => { "backup": "", "changed": false } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:113 Saturday 17 August 2024 19:53:20 -0400 (0:00:00.325) 0:04:31.604 ******* TASK [fedora.linux_system_roles.storage : Show blivet_output] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:119 Saturday 17 August 2024 19:53:20 -0400 (0:00:00.033) 0:04:31.637 ******* ok: [managed_node2] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1" ], "mounts": [ { "fstype": "ext2", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : Set the list of pools for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:128 Saturday 17 August 2024 19:53:20 -0400 (0:00:00.048) 0:04:31.685 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : Set the list of volumes for test verification] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Saturday 17 August 2024 19:53:20 -0400 (0:00:00.047) 0:04:31.733 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Remove obsolete mounts] ************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Saturday 17 August 2024 19:53:20 -0400 (0:00:00.042) 0:04:31.775 ******* changed: [managed_node2] => (item={u'src': u'/dev/mapper/foo-test1', u'state': u'absent', u'fstype': u'ext2', u'path': u'/opt/test1'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "fstype": "ext2", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:159 Saturday 17 August 2024 19:53:20 -0400 (0:00:00.383) 0:04:32.159 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Set up new/current mounts] *********** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:164 Saturday 17 August 2024 19:53:21 -0400 (0:00:00.462) 0:04:32.621 ******* TASK [fedora.linux_system_roles.storage : Manage mount ownership/permissions] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:175 Saturday 17 August 2024 19:53:21 -0400 (0:00:00.034) 0:04:32.656 ******* TASK [fedora.linux_system_roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:187 Saturday 17 August 2024 19:53:21 -0400 (0:00:00.103) 0:04:32.759 ******* ok: [managed_node2] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:195 Saturday 17 August 2024 19:53:21 -0400 (0:00:00.447) 0:04:33.207 ******* ok: [managed_node2] => { "changed": false, "stat": { "atime": 1723936304.5989926, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1723936301.7749832, "dev": 51713, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 917509, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1723936301.7749832, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1736144844", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [fedora.linux_system_roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:200 Saturday 17 August 2024 19:53:22 -0400 (0:00:00.316) 0:04:33.523 ******* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Saturday 17 August 2024 19:53:22 -0400 (0:00:00.034) 0:04:33.557 ******* ok: [managed_node2] TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:315 Saturday 17 August 2024 19:53:22 -0400 (0:00:00.682) 0:04:34.240 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml for managed_node2 TASK [Print out pool information] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:2 Saturday 17 August 2024 19:53:22 -0400 (0:00:00.103) 0:04:34.344 ******* ok: [managed_node2] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_clevis_pin": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "encryption_tang_thumbprint": null, "encryption_tang_url": null, "grow_to_fill": false, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "shared": false, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:7 Saturday 17 August 2024 19:53:22 -0400 (0:00:00.048) 0:04:34.392 ******* skipping: [managed_node2] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:15 Saturday 17 August 2024 19:53:22 -0400 (0:00:00.036) 0:04:34.429 ******* ok: [managed_node2] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "mountpoint": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "ext4", "label": "", "mountpoint": "/", "name": "/dev/xvda1", "size": "250G", "type": "partition", "uuid": "c7b7d6a5-fd01-4b9b-bcca-153eaff9d312" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:20 Saturday 17 August 2024 19:53:23 -0400 (0:00:00.321) 0:04:34.751 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002700", "end": "2024-08-17 19:53:23.493516", "rc": 0, "start": "2024-08-17 19:53:23.490816" } STDOUT: # system_role:storage # # /etc/fstab # Created by anaconda on Thu Jun 20 10:23:46 2024 # # Accessible filesystems, by reference, are maintained under '/dev/disk' # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info # UUID=c7b7d6a5-fd01-4b9b-bcca-153eaff9d312 / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:25 Saturday 17 August 2024 19:53:23 -0400 (0:00:00.314) 0:04:35.065 ******* ok: [managed_node2] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002586", "end": "2024-08-17 19:53:23.811581", "failed_when_result": false, "rc": 0, "start": "2024-08-17 19:53:23.808995" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:34 Saturday 17 August 2024 19:53:23 -0400 (0:00:00.323) 0:04:35.388 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml for managed_node2 TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:5 Saturday 17 August 2024 19:53:23 -0400 (0:00:00.075) 0:04:35.464 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Get VG shared value status] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:18 Saturday 17 August 2024 19:53:23 -0400 (0:00:00.035) 0:04:35.499 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that VG shared value checks out] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:24 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.035) 0:04:35.535 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify pool subset] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool.yml:34 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.035) 0:04:35.570 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:2 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.081) 0:04:35.652 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:8 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.046) 0:04:35.699 ******* TASK [Set pvs lvm length] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:17 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.032) 0:04:35.731 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:22 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.042) 0:04:35.774 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:27 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.045) 0:04:35.819 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:36 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.045) 0:04:35.865 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:41 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.042) 0:04:35.908 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:46 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.044) 0:04:35.952 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:51 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.035) 0:04:35.988 ******* TASK [Check that blivet supports PV grow to fill] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:64 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.031) 0:04:36.020 ******* ok: [managed_node2] => { "changed": false, "rc": 0 } STDOUT: False STDERR: Shared connection to 10.31.46.152 closed. TASK [Verify that PVs fill the whole devices when they should] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:73 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.235) 0:04:36.256 ******* TASK [Check MD RAID] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:83 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.032) 0:04:36.288 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml for managed_node2 TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:8 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.073) 0:04:36.361 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:14 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.036) 0:04:36.397 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:19 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.035) 0:04:36.433 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:24 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.034) 0:04:36.467 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:29 Saturday 17 August 2024 19:53:24 -0400 (0:00:00.033) 0:04:36.501 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:37 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.034) 0:04:36.535 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:46 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:36.571 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:55 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:36.606 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:64 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.034) 0:04:36.641 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:74 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:36.677 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-md.yml:83 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.034) 0:04:36.711 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:86 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:36.747 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml for managed_node2 TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-lvmraid.yml:2 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.075) 0:04:36.822 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml for managed_node2 TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:8 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.075) 0:04:36.898 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:16 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.034) 0:04:36.933 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:20 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.097) 0:04:37.031 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:27 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.036) 0:04:37.067 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:31 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.037) 0:04:37.104 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:37 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:37.139 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-lvmraid.yml:42 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:37.175 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:89 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:37.210 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml for managed_node2 TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-thin.yml:2 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.090) 0:04:37.301 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml for managed_node2 TASK [Get information about thinpool] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:8 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.074) 0:04:37.376 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:16 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:37.411 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:22 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.038) 0:04:37.449 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-thin.yml:26 Saturday 17 August 2024 19:53:25 -0400 (0:00:00.035) 0:04:37.484 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:92 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.035) 0:04:37.520 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml for managed_node2 TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:5 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.080) 0:04:37.601 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:10 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.044) 0:04:37.645 ******* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:17 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.033) 0:04:37.678 ******* TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-encryption.yml:24 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.032) 0:04:37.710 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:95 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.035) 0:04:37.746 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml for managed_node2 TASK [Validate pool member VDO settings] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-members-vdo.yml:2 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.083) 0:04:37.830 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml for managed_node2 TASK [Get information about VDO deduplication] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:8 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.077) 0:04:37.907 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:15 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.036) 0:04:37.943 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:21 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.035) 0:04:37.979 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:27 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.035) 0:04:38.015 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:34 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.039) 0:04:38.055 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:40 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.036) 0:04:38.091 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-member-vdo.yml:46 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.035) 0:04:38.127 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Check Stratis] *********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:98 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.035) 0:04:38.162 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml for managed_node2 TASK [Run 'stratis report'] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:6 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.089) 0:04:38.251 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about Stratis] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:11 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.050) 0:04:38.302 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the pools was created] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:15 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.034) 0:04:38.337 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that encryption is correctly set] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:25 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.034) 0:04:38.371 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that Clevis/Tang encryption is correctly set] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:34 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.034) 0:04:38.405 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-pool-stratis.yml:44 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.036) 0:04:38.441 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_stratis_report": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-members.yml:101 Saturday 17 August 2024 19:53:26 -0400 (0:00:00.038) 0:04:38.479 ******* ok: [managed_node2] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-pool-volumes.yml:3 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:38.515 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml for managed_node2 TASK [Set storage volume test variables] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:2 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.068) 0:04:38.584 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:19 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.043) 0:04:38.627 ******* included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml for managed_node2 included: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml for managed_node2 TASK [Get expected mount device based on device type] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:7 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.178) 0:04:38.805 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:11 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.040) 0:04:38.845 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_mount_expected_mount_point": "/opt/test1", "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:19 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.043) 0:04:38.889 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:28 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:38.925 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory user] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:36 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.029) 0:04:38.954 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:42 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:38.989 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:48 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.037) 0:04:39.026 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:57 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:39.062 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:63 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.036) 0:04:39.098 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:69 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.034) 0:04:39.133 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-mount.yml:79 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.034) 0:04:39.168 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_expected_mount_point": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:2 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:39.203 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:17 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.135) 0:04:39.339 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:24 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:39.374 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:33 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.043) 0:04:39.418 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fingerprint] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:45 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.035) 0:04:39.454 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Clean up variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fstab.yml:51 Saturday 17 August 2024 19:53:27 -0400 (0:00:00.041) 0:04:39.495 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:6 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.036) 0:04:39.532 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-fs.yml:14 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.041) 0:04:39.574 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:3 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.038) 0:04:39.613 ******* ok: [managed_node2] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:9 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.327) 0:04:39.941 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:16 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.037) 0:04:39.978 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:23 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.047) 0:04:40.025 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:29 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.030) 0:04:40.055 ******* ok: [managed_node2] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:33 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.039) 0:04:40.095 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-device.yml:38 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.036) 0:04:40.131 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:3 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.033) 0:04:40.164 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:10 Saturday 17 August 2024 19:53:28 -0400 (0:00:00.036) 0:04:40.201 ******* ok: [managed_node2] => { "changed": false, "rc": 0, "results": [ "cryptsetup-2.0.3-6.el7.x86_64 providing cryptsetup is already installed" ] } lsrpackages: cryptsetup TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:16 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.507) 0:04:40.708 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:22 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:40.745 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:29 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:40.782 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:40 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.030) 0:04:40.813 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:46 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.038) 0:04:40.851 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:51 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:40.888 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:63 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.030) 0:04:40.918 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:75 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.030) 0:04:40.949 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:87 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.030) 0:04:40.979 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:93 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.044) 0:04:41.024 ******* ok: [managed_node2] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:100 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.045) 0:04:41.070 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:108 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:41.106 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:116 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:41.143 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-encryption.yml:124 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:41.180 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:8 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:41.216 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:14 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.037) 0:04:41.254 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:19 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.040) 0:04:41.294 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:24 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.037) 0:04:41.331 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:29 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.041) 0:04:41.373 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:37 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.047) 0:04:41.420 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:46 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.037) 0:04:41.458 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:54 Saturday 17 August 2024 19:53:29 -0400 (0:00:00.036) 0:04:41.494 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:62 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.039) 0:04:41.534 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-md.yml:70 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.037) 0:04:41.572 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:3 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.609 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:11 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.645 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:20 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.035) 0:04:41.681 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:28 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.717 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:32 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.043) 0:04:41.761 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:46 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.797 ******* skipping: [managed_node2] => {} TASK [Show test blockinfo] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:50 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.037) 0:04:41.834 ******* skipping: [managed_node2] => {} TASK [Show test pool size] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:54 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.871 ******* skipping: [managed_node2] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:58 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.907 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:67 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:41.943 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:71 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.039) 0:04:41.982 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:76 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.037) 0:04:42.020 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:82 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.037) 0:04:42.057 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:86 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:42.093 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:91 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.036) 0:04:42.130 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:96 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.035) 0:04:42.165 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:101 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.039) 0:04:42.204 ******* skipping: [managed_node2] => {} TASK [Show volume thin pool size] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:105 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.035) 0:04:42.239 ******* skipping: [managed_node2] => {} TASK [Show test volume size] *************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:109 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.035) 0:04:42.275 ******* skipping: [managed_node2] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:113 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.035) 0:04:42.310 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:120 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.034) 0:04:42.345 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:127 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.033) 0:04:42.379 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:131 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.045) 0:04:42.424 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:137 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.039) 0:04:42.463 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:143 Saturday 17 August 2024 19:53:30 -0400 (0:00:00.037) 0:04:42.501 ******* ok: [managed_node2] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:147 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.041) 0:04:42.542 ******* ok: [managed_node2] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-size.yml:151 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.039) 0:04:42.582 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:5 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.035) 0:04:42.618 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:13 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.038) 0:04:42.657 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:17 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.099) 0:04:42.756 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:24 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.037) 0:04:42.793 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:31 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.036) 0:04:42.830 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:37 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.036) 0:04:42.867 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume-cache.yml:42 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.036) 0:04:42.904 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/test-verify-volume.yml:25 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.036) 0:04:42.940 ******* ok: [managed_node2] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:44 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.039) 0:04:42.980 ******* TASK [Clean up variable namespace] ********************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/verify-role-results.yml:54 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.033) 0:04:43.013 ******* ok: [managed_node2] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Gather package facts] **************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:318 Saturday 17 August 2024 19:53:31 -0400 (0:00:00.035) 0:04:43.049 ******* ok: [managed_node2] => { "ansible_facts": { "packages": { "NetworkManager": [ { "arch": "x86_64", "epoch": 1, "name": "NetworkManager", "release": "2.el7_9", "source": "rpm", "version": "1.18.8" } ], "NetworkManager-libnm": [ { "arch": "x86_64", "epoch": 1, "name": "NetworkManager-libnm", "release": "2.el7_9", "source": "rpm", "version": "1.18.8" } ], "NetworkManager-team": [ { "arch": "x86_64", "epoch": 1, "name": "NetworkManager-team", "release": "2.el7_9", "source": "rpm", "version": "1.18.8" } ], "NetworkManager-tui": [ { "arch": "x86_64", "epoch": 1, "name": "NetworkManager-tui", "release": "2.el7_9", "source": "rpm", "version": "1.18.8" } ], "PyYAML": [ { "arch": "x86_64", "epoch": null, "name": "PyYAML", "release": "11.el7", "source": "rpm", "version": "3.10" } ], "acl": [ { "arch": "x86_64", "epoch": null, "name": "acl", "release": "15.el7", "source": "rpm", "version": "2.2.51" } ], "aic94xx-firmware": [ { "arch": "noarch", "epoch": null, "name": "aic94xx-firmware", "release": "6.el7", "source": "rpm", "version": "30" } ], "alsa-firmware": [ { "arch": "noarch", "epoch": null, "name": "alsa-firmware", "release": "2.el7", "source": "rpm", "version": "1.0.28" } ], "alsa-lib": [ { "arch": "x86_64", "epoch": null, "name": "alsa-lib", "release": "1.el7", "source": "rpm", "version": "1.1.8" } ], "alsa-tools-firmware": [ { "arch": "x86_64", "epoch": null, "name": "alsa-tools-firmware", "release": "1.el7", "source": "rpm", "version": "1.1.0" } ], "aspell": [ { "arch": "x86_64", "epoch": 12, "name": "aspell", "release": "9.el7", "source": "rpm", "version": "0.60.6.1" } ], "audit": [ { "arch": "x86_64", "epoch": null, "name": "audit", "release": "4.el7", "source": "rpm", "version": "2.8.5" } ], "audit-libs": [ { "arch": "x86_64", "epoch": null, "name": "audit-libs", "release": "4.el7", "source": "rpm", "version": "2.8.5" } ], "audit-libs-python": [ { "arch": "x86_64", "epoch": null, "name": "audit-libs-python", "release": "4.el7", "source": "rpm", "version": "2.8.5" } ], "authconfig": [ { "arch": "x86_64", "epoch": null, "name": "authconfig", "release": "30.el7", "source": "rpm", "version": "6.2.8" } ], "avahi-libs": [ { "arch": "x86_64", "epoch": null, "name": "avahi-libs", "release": "20.el7", "source": "rpm", "version": "0.6.31" } ], "basesystem": [ { "arch": "noarch", "epoch": null, "name": "basesystem", "release": "7.el7.centos", "source": "rpm", "version": "10.0" } ], "bash": [ { "arch": "x86_64", "epoch": null, "name": "bash", "release": "35.el7_9", "source": "rpm", "version": "4.2.46" } ], "bc": [ { "arch": "x86_64", "epoch": null, "name": "bc", "release": "13.el7", "source": "rpm", "version": "1.06.95" } ], "beakerlib": [ { "arch": "noarch", "epoch": null, "name": "beakerlib", "release": "1.el7bkr", "source": "rpm", "version": "1.29.3" } ], "beakerlib-redhat": [ { "arch": "noarch", "epoch": null, "name": "beakerlib-redhat", "release": "33.el7bkr", "source": "rpm", "version": "1" } ], "bind-export-libs": [ { "arch": "x86_64", "epoch": 32, "name": "bind-export-libs", "release": "26.P2.el7_9.16", "source": "rpm", "version": "9.11.4" } ], "binutils": [ { "arch": "x86_64", "epoch": null, "name": "binutils", "release": "44.base.el7_9.1", "source": "rpm", "version": "2.27" } ], "biosdevname": [ { "arch": "x86_64", "epoch": null, "name": "biosdevname", "release": "2.el7", "source": "rpm", "version": "0.7.3" } ], "blivet3-data": [ { "arch": "noarch", "epoch": 1, "name": "blivet3-data", "release": "3.el7", "source": "rpm", "version": "3.1.3" } ], "boost-date-time": [ { "arch": "x86_64", "epoch": null, "name": "boost-date-time", "release": "28.el7", "source": "rpm", "version": "1.53.0" } ], "boost-system": [ { "arch": "x86_64", "epoch": null, "name": "boost-system", "release": "28.el7", "source": "rpm", "version": "1.53.0" } ], "boost-thread": [ { "arch": "x86_64", "epoch": null, "name": "boost-thread", "release": "28.el7", "source": "rpm", "version": "1.53.0" } ], "btrfs-progs": [ { "arch": "x86_64", "epoch": null, "name": "btrfs-progs", "release": "1.el7", "source": "rpm", "version": "4.9.1" } ], "bzip2-libs": [ { "arch": "x86_64", "epoch": null, "name": "bzip2-libs", "release": "13.el7", "source": "rpm", "version": "1.0.6" } ], "ca-certificates": [ { "arch": "noarch", "epoch": null, "name": "ca-certificates", "release": "72.el7_9", "source": "rpm", "version": "2023.2.60_v7.0.306" } ], "centos-logos": [ { "arch": "noarch", "epoch": null, "name": "centos-logos", "release": "3.el7.centos", "source": "rpm", "version": "70.0.6" } ], "centos-release": [ { "arch": "x86_64", "epoch": null, "name": "centos-release", "release": "9.2009.2.el7.centos", "source": "rpm", "version": "7" } ], "checkpolicy": [ { "arch": "x86_64", "epoch": null, "name": "checkpolicy", "release": "8.el7", "source": "rpm", "version": "2.5" } ], "chkconfig": [ { "arch": "x86_64", "epoch": null, "name": "chkconfig", "release": "1.el7", "source": "rpm", "version": "1.7.6" } ], "chrony": [ { "arch": "x86_64", "epoch": null, "name": "chrony", "release": "1.el7", "source": "rpm", "version": "3.4" } ], "cloud-init": [ { "arch": "x86_64", "epoch": null, "name": "cloud-init", "release": "24.el7", "source": "rpm", "version": "0.7.9" } ], "cloud-utils-growpart": [ { "arch": "noarch", "epoch": null, "name": "cloud-utils-growpart", "release": "5.el7", "source": "rpm", "version": "0.29" } ], "coreutils": [ { "arch": "x86_64", "epoch": null, "name": "coreutils", "release": "24.el7_9.2", "source": "rpm", "version": "8.22" } ], "cpio": [ { "arch": "x86_64", "epoch": null, "name": "cpio", "release": "28.el7", "source": "rpm", "version": "2.11" } ], "cpp": [ { "arch": "x86_64", "epoch": null, "name": "cpp", "release": "44.el7", "source": "rpm", "version": "4.8.5" } ], "cracklib": [ { "arch": "x86_64", "epoch": null, "name": "cracklib", "release": "11.el7", "source": "rpm", "version": "2.9.0" } ], "cracklib-dicts": [ { "arch": "x86_64", "epoch": null, "name": "cracklib-dicts", "release": "11.el7", "source": "rpm", "version": "2.9.0" } ], "cronie": [ { "arch": "x86_64", "epoch": null, "name": "cronie", "release": "25.el7_9", "source": "rpm", "version": "1.4.11" } ], "cronie-anacron": [ { "arch": "x86_64", "epoch": null, "name": "cronie-anacron", "release": "25.el7_9", "source": "rpm", "version": "1.4.11" } ], "crontabs": [ { "arch": "noarch", "epoch": null, "name": "crontabs", "release": "6.20121102git.el7", "source": "rpm", "version": "1.11" } ], "cryptsetup": [ { "arch": "x86_64", "epoch": null, "name": "cryptsetup", "release": "6.el7", "source": "rpm", "version": "2.0.3" } ], "cryptsetup-libs": [ { "arch": "x86_64", "epoch": null, "name": "cryptsetup-libs", "release": "6.el7", "source": "rpm", "version": "2.0.3" } ], "curl": [ { "arch": "x86_64", "epoch": null, "name": "curl", "release": "59.el7_9.2", "source": "rpm", "version": "7.29.0" } ], "cyrus-sasl-lib": [ { "arch": "x86_64", "epoch": null, "name": "cyrus-sasl-lib", "release": "24.el7_9", "source": "rpm", "version": "2.1.26" } ], "dbus": [ { "arch": "x86_64", "epoch": 1, "name": "dbus", "release": "15.el7", "source": "rpm", "version": "1.10.24" } ], "dbus-glib": [ { "arch": "x86_64", "epoch": null, "name": "dbus-glib", "release": "7.el7", "source": "rpm", "version": "0.100" } ], "dbus-libs": [ { "arch": "x86_64", "epoch": 1, "name": "dbus-libs", "release": "15.el7", "source": "rpm", "version": "1.10.24" } ], "dbus-python": [ { "arch": "x86_64", "epoch": null, "name": "dbus-python", "release": "9.el7", "source": "rpm", "version": "1.1.1" } ], "device-mapper": [ { "arch": "x86_64", "epoch": 7, "name": "device-mapper", "release": "6.el7_9.5", "source": "rpm", "version": "1.02.170" } ], "device-mapper-event": [ { "arch": "x86_64", "epoch": 7, "name": "device-mapper-event", "release": "6.el7_9.5", "source": "rpm", "version": "1.02.170" } ], "device-mapper-event-libs": [ { "arch": "x86_64", "epoch": 7, "name": "device-mapper-event-libs", "release": "6.el7_9.5", "source": "rpm", "version": "1.02.170" } ], "device-mapper-libs": [ { "arch": "x86_64", "epoch": 7, "name": "device-mapper-libs", "release": "6.el7_9.5", "source": "rpm", "version": "1.02.170" } ], "device-mapper-persistent-data": [ { "arch": "x86_64", "epoch": null, "name": "device-mapper-persistent-data", "release": "3.el7_9.2", "source": "rpm", "version": "0.8.5" } ], "dhclient": [ { "arch": "x86_64", "epoch": 12, "name": "dhclient", "release": "83.el7.centos.2", "source": "rpm", "version": "4.2.5" } ], "dhcp-common": [ { "arch": "x86_64", "epoch": 12, "name": "dhcp-common", "release": "83.el7.centos.2", "source": "rpm", "version": "4.2.5" } ], "dhcp-libs": [ { "arch": "x86_64", "epoch": 12, "name": "dhcp-libs", "release": "83.el7.centos.2", "source": "rpm", "version": "4.2.5" } ], "diffutils": [ { "arch": "x86_64", "epoch": null, "name": "diffutils", "release": "6.el7_9", "source": "rpm", "version": "3.3" } ], "dmidecode": [ { "arch": "x86_64", "epoch": 1, "name": "dmidecode", "release": "5.el7_9.1", "source": "rpm", "version": "3.2" } ], "dmraid": [ { "arch": "x86_64", "epoch": null, "name": "dmraid", "release": "28.el7", "source": "rpm", "version": "1.0.0.rc16" } ], "dmraid-events": [ { "arch": "x86_64", "epoch": null, "name": "dmraid-events", "release": "28.el7", "source": "rpm", "version": "1.0.0.rc16" } ], "dracut": [ { "arch": "x86_64", "epoch": null, "name": "dracut", "release": "572.el7", "source": "rpm", "version": "033" } ], "dracut-config-rescue": [ { "arch": "x86_64", "epoch": null, "name": "dracut-config-rescue", "release": "572.el7", "source": "rpm", "version": "033" } ], "dracut-network": [ { "arch": "x86_64", "epoch": null, "name": "dracut-network", "release": "572.el7", "source": "rpm", "version": "033" } ], "dyninst": [ { "arch": "x86_64", "epoch": null, "name": "dyninst", "release": "3.el7", "source": "rpm", "version": "9.3.1" } ], "e2fsprogs": [ { "arch": "x86_64", "epoch": null, "name": "e2fsprogs", "release": "19.el7", "source": "rpm", "version": "1.42.9" } ], "e2fsprogs-libs": [ { "arch": "x86_64", "epoch": null, "name": "e2fsprogs-libs", "release": "19.el7", "source": "rpm", "version": "1.42.9" } ], "ebtables": [ { "arch": "x86_64", "epoch": null, "name": "ebtables", "release": "16.el7", "source": "rpm", "version": "2.0.10" } ], "efivar-libs": [ { "arch": "x86_64", "epoch": null, "name": "efivar-libs", "release": "12.el7", "source": "rpm", "version": "36" } ], "elfutils-default-yama-scope": [ { "arch": "noarch", "epoch": null, "name": "elfutils-default-yama-scope", "release": "5.el7", "source": "rpm", "version": "0.176" } ], "elfutils-libelf": [ { "arch": "x86_64", "epoch": null, "name": "elfutils-libelf", "release": "5.el7", "source": "rpm", "version": "0.176" } ], "elfutils-libs": [ { "arch": "x86_64", "epoch": null, "name": "elfutils-libs", "release": "5.el7", "source": "rpm", "version": "0.176" } ], "epel-release": [ { "arch": "noarch", "epoch": null, "name": "epel-release", "release": "14", "source": "rpm", "version": "7" } ], "ethtool": [ { "arch": "x86_64", "epoch": 2, "name": "ethtool", "release": "10.el7", "source": "rpm", "version": "4.8" } ], "expat": [ { "arch": "x86_64", "epoch": null, "name": "expat", "release": "15.el7_9", "source": "rpm", "version": "2.1.0" } ], "file": [ { "arch": "x86_64", "epoch": null, "name": "file", "release": "37.el7", "source": "rpm", "version": "5.11" } ], "file-libs": [ { "arch": "x86_64", "epoch": null, "name": "file-libs", "release": "37.el7", "source": "rpm", "version": "5.11" } ], "filesystem": [ { "arch": "x86_64", "epoch": null, "name": "filesystem", "release": "25.el7", "source": "rpm", "version": "3.2" } ], "findutils": [ { "arch": "x86_64", "epoch": 1, "name": "findutils", "release": "6.el7", "source": "rpm", "version": "4.5.11" } ], "fipscheck": [ { "arch": "x86_64", "epoch": null, "name": "fipscheck", "release": "6.el7", "source": "rpm", "version": "1.4.1" } ], "fipscheck-lib": [ { "arch": "x86_64", "epoch": null, "name": "fipscheck-lib", "release": "6.el7", "source": "rpm", "version": "1.4.1" } ], "firewalld": [ { "arch": "noarch", "epoch": null, "name": "firewalld", "release": "13.el7_9", "source": "rpm", "version": "0.6.3" } ], "firewalld-filesystem": [ { "arch": "noarch", "epoch": null, "name": "firewalld-filesystem", "release": "13.el7_9", "source": "rpm", "version": "0.6.3" } ], "freetype": [ { "arch": "x86_64", "epoch": null, "name": "freetype", "release": "14.el7_9.1", "source": "rpm", "version": "2.8" } ], "fxload": [ { "arch": "x86_64", "epoch": null, "name": "fxload", "release": "16.el7", "source": "rpm", "version": "2002_04_11" } ], "gawk": [ { "arch": "x86_64", "epoch": null, "name": "gawk", "release": "4.el7_3.1", "source": "rpm", "version": "4.0.2" } ], "gcc": [ { "arch": "x86_64", "epoch": null, "name": "gcc", "release": "44.el7", "source": "rpm", "version": "4.8.5" } ], "gdbm": [ { "arch": "x86_64", "epoch": null, "name": "gdbm", "release": "8.el7", "source": "rpm", "version": "1.10" } ], "gettext": [ { "arch": "x86_64", "epoch": null, "name": "gettext", "release": "3.el7", "source": "rpm", "version": "0.19.8.1" } ], "gettext-libs": [ { "arch": "x86_64", "epoch": null, "name": "gettext-libs", "release": "3.el7", "source": "rpm", "version": "0.19.8.1" } ], "git": [ { "arch": "x86_64", "epoch": null, "name": "git", "release": "25.el7_9", "source": "rpm", "version": "1.8.3.1" } ], "glib2": [ { "arch": "x86_64", "epoch": null, "name": "glib2", "release": "9.el7_9", "source": "rpm", "version": "2.56.1" } ], "glibc": [ { "arch": "x86_64", "epoch": null, "name": "glibc", "release": "326.el7_9.3", "source": "rpm", "version": "2.17" } ], "glibc-common": [ { "arch": "x86_64", "epoch": null, "name": "glibc-common", "release": "326.el7_9.3", "source": "rpm", "version": "2.17" } ], "glibc-devel": [ { "arch": "x86_64", "epoch": null, "name": "glibc-devel", "release": "326.el7_9.3", "source": "rpm", "version": "2.17" } ], "glibc-headers": [ { "arch": "x86_64", "epoch": null, "name": "glibc-headers", "release": "326.el7_9.3", "source": "rpm", "version": "2.17" } ], "gmp": [ { "arch": "x86_64", "epoch": 1, "name": "gmp", "release": "15.el7", "source": "rpm", "version": "6.0.0" } ], "gnupg2": [ { "arch": "x86_64", "epoch": null, "name": "gnupg2", "release": "5.el7_5", "source": "rpm", "version": "2.0.22" } ], "gobject-introspection": [ { "arch": "x86_64", "epoch": null, "name": "gobject-introspection", "release": "1.el7", "source": "rpm", "version": "1.56.1" } ], "gpg-pubkey": [ { "arch": null, "epoch": null, "name": "gpg-pubkey", "release": "52ae6884", "source": "rpm", "version": "352c64e5" }, { "arch": null, "epoch": null, "name": "gpg-pubkey", "release": "53a7ff4b", "source": "rpm", "version": "f4a80eb5" } ], "gpgme": [ { "arch": "x86_64", "epoch": null, "name": "gpgme", "release": "5.el7", "source": "rpm", "version": "1.3.2" } ], "gpm-libs": [ { "arch": "x86_64", "epoch": null, "name": "gpm-libs", "release": "6.el7", "source": "rpm", "version": "1.20.7" } ], "grep": [ { "arch": "x86_64", "epoch": null, "name": "grep", "release": "3.el7", "source": "rpm", "version": "2.20" } ], "groff-base": [ { "arch": "x86_64", "epoch": null, "name": "groff-base", "release": "8.el7", "source": "rpm", "version": "1.22.2" } ], "grub2": [ { "arch": "x86_64", "epoch": 1, "name": "grub2", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grub2-common": [ { "arch": "noarch", "epoch": 1, "name": "grub2-common", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grub2-pc": [ { "arch": "x86_64", "epoch": 1, "name": "grub2-pc", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grub2-pc-modules": [ { "arch": "noarch", "epoch": 1, "name": "grub2-pc-modules", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grub2-tools": [ { "arch": "x86_64", "epoch": 1, "name": "grub2-tools", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grub2-tools-extra": [ { "arch": "x86_64", "epoch": 1, "name": "grub2-tools-extra", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grub2-tools-minimal": [ { "arch": "x86_64", "epoch": 1, "name": "grub2-tools-minimal", "release": "0.87.0.2.el7.centos.14", "source": "rpm", "version": "2.02" } ], "grubby": [ { "arch": "x86_64", "epoch": null, "name": "grubby", "release": "26.el7", "source": "rpm", "version": "8.28" } ], "gssproxy": [ { "arch": "x86_64", "epoch": null, "name": "gssproxy", "release": "30.el7_9", "source": "rpm", "version": "0.7.0" } ], "gzip": [ { "arch": "x86_64", "epoch": null, "name": "gzip", "release": "11.el7_9", "source": "rpm", "version": "1.5" } ], "hardlink": [ { "arch": "x86_64", "epoch": 1, "name": "hardlink", "release": "19.el7", "source": "rpm", "version": "1.0" } ], "hostname": [ { "arch": "x86_64", "epoch": null, "name": "hostname", "release": "3.el7_7.1", "source": "rpm", "version": "3.13" } ], "hwdata": [ { "arch": "x86_64", "epoch": null, "name": "hwdata", "release": "9.7.el7", "source": "rpm", "version": "0.252" } ], "info": [ { "arch": "x86_64", "epoch": null, "name": "info", "release": "5.el7", "source": "rpm", "version": "5.1" } ], "initscripts": [ { "arch": "x86_64", "epoch": null, "name": "initscripts", "release": "1.el7_9.1", "source": "rpm", "version": "9.49.53" } ], "iproute": [ { "arch": "x86_64", "epoch": null, "name": "iproute", "release": "30.el7", "source": "rpm", "version": "4.11.0" } ], "iprutils": [ { "arch": "x86_64", "epoch": null, "name": "iprutils", "release": "3.el7_7", "source": "rpm", "version": "2.4.17.1" } ], "ipset": [ { "arch": "x86_64", "epoch": null, "name": "ipset", "release": "1.el7", "source": "rpm", "version": "7.1" } ], "ipset-libs": [ { "arch": "x86_64", "epoch": null, "name": "ipset-libs", "release": "1.el7", "source": "rpm", "version": "7.1" } ], "iptables": [ { "arch": "x86_64", "epoch": null, "name": "iptables", "release": "35.el7", "source": "rpm", "version": "1.4.21" } ], "iputils": [ { "arch": "x86_64", "epoch": null, "name": "iputils", "release": "10.el7", "source": "rpm", "version": "20160308" } ], "irqbalance": [ { "arch": "x86_64", "epoch": 3, "name": "irqbalance", "release": "12.el7", "source": "rpm", "version": "1.0.7" } ], "ivtv-firmware": [ { "arch": "noarch", "epoch": 2, "name": "ivtv-firmware", "release": "26.el7", "source": "rpm", "version": "20080701" } ], "iwl100-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl100-firmware", "release": "82.el7_9", "source": "rpm", "version": "39.31.5.1" } ], "iwl1000-firmware": [ { "arch": "noarch", "epoch": 1, "name": "iwl1000-firmware", "release": "82.el7_9", "source": "rpm", "version": "39.31.5.1" } ], "iwl105-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl105-firmware", "release": "82.el7_9", "source": "rpm", "version": "18.168.6.1" } ], "iwl135-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl135-firmware", "release": "82.el7_9", "source": "rpm", "version": "18.168.6.1" } ], "iwl2000-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl2000-firmware", "release": "82.el7_9", "source": "rpm", "version": "18.168.6.1" } ], "iwl2030-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl2030-firmware", "release": "82.el7_9", "source": "rpm", "version": "18.168.6.1" } ], "iwl3160-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl3160-firmware", "release": "82.el7_9", "source": "rpm", "version": "25.30.13.0" } ], "iwl3945-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl3945-firmware", "release": "82.el7_9", "source": "rpm", "version": "15.32.2.9" } ], "iwl4965-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl4965-firmware", "release": "82.el7_9", "source": "rpm", "version": "228.61.2.24" } ], "iwl5000-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl5000-firmware", "release": "82.el7_9", "source": "rpm", "version": "8.83.5.1_1" } ], "iwl5150-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl5150-firmware", "release": "82.el7_9", "source": "rpm", "version": "8.24.2.2" } ], "iwl6000-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl6000-firmware", "release": "82.el7_9", "source": "rpm", "version": "9.221.4.1" } ], "iwl6000g2a-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl6000g2a-firmware", "release": "82.el7_9", "source": "rpm", "version": "18.168.6.1" } ], "iwl6000g2b-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl6000g2b-firmware", "release": "82.el7_9", "source": "rpm", "version": "18.168.6.1" } ], "iwl6050-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl6050-firmware", "release": "82.el7_9", "source": "rpm", "version": "41.28.5.1" } ], "iwl7260-firmware": [ { "arch": "noarch", "epoch": null, "name": "iwl7260-firmware", "release": "82.el7_9", "source": "rpm", "version": "25.30.13.0" } ], "jansson": [ { "arch": "x86_64", "epoch": null, "name": "jansson", "release": "1.el7", "source": "rpm", "version": "2.10" } ], "json-c": [ { "arch": "x86_64", "epoch": null, "name": "json-c", "release": "4.el7_0", "source": "rpm", "version": "0.11" } ], "kbd": [ { "arch": "x86_64", "epoch": null, "name": "kbd", "release": "16.el7_9", "source": "rpm", "version": "1.15.5" } ], "kbd-legacy": [ { "arch": "noarch", "epoch": null, "name": "kbd-legacy", "release": "16.el7_9", "source": "rpm", "version": "1.15.5" } ], "kbd-misc": [ { "arch": "noarch", "epoch": null, "name": "kbd-misc", "release": "16.el7_9", "source": "rpm", "version": "1.15.5" } ], "kernel": [ { "arch": "x86_64", "epoch": null, "name": "kernel", "release": "1160.119.1.el7", "source": "rpm", "version": "3.10.0" }, { "arch": "x86_64", "epoch": null, "name": "kernel", "release": "1160.el7", "source": "rpm", "version": "3.10.0" } ], "kernel-debug-devel": [ { "arch": "x86_64", "epoch": null, "name": "kernel-debug-devel", "release": "1160.119.1.el7", "source": "rpm", "version": "3.10.0" } ], "kernel-headers": [ { "arch": "x86_64", "epoch": null, "name": "kernel-headers", "release": "1160.119.1.el7", "source": "rpm", "version": "3.10.0" } ], "kernel-tools": [ { "arch": "x86_64", "epoch": null, "name": "kernel-tools", "release": "1160.119.1.el7", "source": "rpm", "version": "3.10.0" } ], "kernel-tools-libs": [ { "arch": "x86_64", "epoch": null, "name": "kernel-tools-libs", "release": "1160.119.1.el7", "source": "rpm", "version": "3.10.0" } ], "kexec-tools": [ { "arch": "x86_64", "epoch": null, "name": "kexec-tools", "release": "51.el7_9.3", "source": "rpm", "version": "2.0.15" } ], "keyutils": [ { "arch": "x86_64", "epoch": null, "name": "keyutils", "release": "3.el7", "source": "rpm", "version": "1.5.8" } ], "keyutils-libs": [ { "arch": "x86_64", "epoch": null, "name": "keyutils-libs", "release": "3.el7", "source": "rpm", "version": "1.5.8" } ], "kmod": [ { "arch": "x86_64", "epoch": null, "name": "kmod", "release": "28.el7", "source": "rpm", "version": "20" } ], "kmod-libs": [ { "arch": "x86_64", "epoch": null, "name": "kmod-libs", "release": "28.el7", "source": "rpm", "version": "20" } ], "kpartx": [ { "arch": "x86_64", "epoch": null, "name": "kpartx", "release": "136.el7_9", "source": "rpm", "version": "0.4.9" } ], "krb5-libs": [ { "arch": "x86_64", "epoch": null, "name": "krb5-libs", "release": "55.el7_9", "source": "rpm", "version": "1.15.1" } ], "less": [ { "arch": "x86_64", "epoch": null, "name": "less", "release": "10.el7_9", "source": "rpm", "version": "458" } ], "libacl": [ { "arch": "x86_64", "epoch": null, "name": "libacl", "release": "15.el7", "source": "rpm", "version": "2.2.51" } ], "libaio": [ { "arch": "x86_64", "epoch": null, "name": "libaio", "release": "13.el7", "source": "rpm", "version": "0.3.109" } ], "libassuan": [ { "arch": "x86_64", "epoch": null, "name": "libassuan", "release": "3.el7", "source": "rpm", "version": "2.1.0" } ], "libattr": [ { "arch": "x86_64", "epoch": null, "name": "libattr", "release": "13.el7", "source": "rpm", "version": "2.4.46" } ], "libbasicobjects": [ { "arch": "x86_64", "epoch": null, "name": "libbasicobjects", "release": "32.el7", "source": "rpm", "version": "0.1.1" } ], "libblkid": [ { "arch": "x86_64", "epoch": null, "name": "libblkid", "release": "65.el7_9.1", "source": "rpm", "version": "2.23.2" } ], "libblockdev": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libblockdev-crypto": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev-crypto", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libblockdev-dm": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev-dm", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libblockdev-lvm": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev-lvm", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libblockdev-mdraid": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev-mdraid", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libblockdev-swap": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev-swap", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libblockdev-utils": [ { "arch": "x86_64", "epoch": null, "name": "libblockdev-utils", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "libbytesize": [ { "arch": "x86_64", "epoch": null, "name": "libbytesize", "release": "1.el7", "source": "rpm", "version": "1.2" } ], "libcap": [ { "arch": "x86_64", "epoch": null, "name": "libcap", "release": "11.el7", "source": "rpm", "version": "2.22" } ], "libcap-ng": [ { "arch": "x86_64", "epoch": null, "name": "libcap-ng", "release": "4.el7", "source": "rpm", "version": "0.7.5" } ], "libcgroup": [ { "arch": "x86_64", "epoch": null, "name": "libcgroup", "release": "21.el7", "source": "rpm", "version": "0.41" } ], "libcollection": [ { "arch": "x86_64", "epoch": null, "name": "libcollection", "release": "32.el7", "source": "rpm", "version": "0.7.0" } ], "libcom_err": [ { "arch": "x86_64", "epoch": null, "name": "libcom_err", "release": "19.el7", "source": "rpm", "version": "1.42.9" } ], "libcroco": [ { "arch": "x86_64", "epoch": null, "name": "libcroco", "release": "6.el7_9", "source": "rpm", "version": "0.6.12" } ], "libcurl": [ { "arch": "x86_64", "epoch": null, "name": "libcurl", "release": "59.el7_9.2", "source": "rpm", "version": "7.29.0" } ], "libdaemon": [ { "arch": "x86_64", "epoch": null, "name": "libdaemon", "release": "7.el7", "source": "rpm", "version": "0.14" } ], "libdb": [ { "arch": "x86_64", "epoch": null, "name": "libdb", "release": "25.el7", "source": "rpm", "version": "5.3.21" } ], "libdb-utils": [ { "arch": "x86_64", "epoch": null, "name": "libdb-utils", "release": "25.el7", "source": "rpm", "version": "5.3.21" } ], "libdnf": [ { "arch": "x86_64", "epoch": null, "name": "libdnf", "release": "2.el7_9", "source": "rpm", "version": "0.22.5" } ], "libdrm": [ { "arch": "x86_64", "epoch": null, "name": "libdrm", "release": "2.el7", "source": "rpm", "version": "2.4.97" } ], "libdwarf": [ { "arch": "x86_64", "epoch": null, "name": "libdwarf", "release": "4.el7", "source": "rpm", "version": "20130207" } ], "libedit": [ { "arch": "x86_64", "epoch": null, "name": "libedit", "release": "12.20121213cvs.el7", "source": "rpm", "version": "3.0" } ], "libestr": [ { "arch": "x86_64", "epoch": null, "name": "libestr", "release": "2.el7", "source": "rpm", "version": "0.1.9" } ], "libevent": [ { "arch": "x86_64", "epoch": null, "name": "libevent", "release": "4.el7", "source": "rpm", "version": "2.0.21" } ], "libfastjson": [ { "arch": "x86_64", "epoch": null, "name": "libfastjson", "release": "3.el7", "source": "rpm", "version": "0.99.4" } ], "libffi": [ { "arch": "x86_64", "epoch": null, "name": "libffi", "release": "19.el7", "source": "rpm", "version": "3.0.13" } ], "libgcc": [ { "arch": "x86_64", "epoch": null, "name": "libgcc", "release": "44.el7", "source": "rpm", "version": "4.8.5" } ], "libgcrypt": [ { "arch": "x86_64", "epoch": null, "name": "libgcrypt", "release": "14.el7", "source": "rpm", "version": "1.5.3" } ], "libgomp": [ { "arch": "x86_64", "epoch": null, "name": "libgomp", "release": "44.el7", "source": "rpm", "version": "4.8.5" } ], "libgpg-error": [ { "arch": "x86_64", "epoch": null, "name": "libgpg-error", "release": "3.el7", "source": "rpm", "version": "1.12" } ], "libidn": [ { "arch": "x86_64", "epoch": null, "name": "libidn", "release": "4.el7", "source": "rpm", "version": "1.28" } ], "libini_config": [ { "arch": "x86_64", "epoch": null, "name": "libini_config", "release": "32.el7", "source": "rpm", "version": "1.3.1" } ], "libmnl": [ { "arch": "x86_64", "epoch": null, "name": "libmnl", "release": "7.el7", "source": "rpm", "version": "1.0.3" } ], "libmodulemd": [ { "arch": "x86_64", "epoch": null, "name": "libmodulemd", "release": "1.el7", "source": "rpm", "version": "1.6.3" } ], "libmount": [ { "arch": "x86_64", "epoch": null, "name": "libmount", "release": "65.el7_9.1", "source": "rpm", "version": "2.23.2" } ], "libmpc": [ { "arch": "x86_64", "epoch": null, "name": "libmpc", "release": "3.el7", "source": "rpm", "version": "1.0.1" } ], "libndp": [ { "arch": "x86_64", "epoch": null, "name": "libndp", "release": "9.el7", "source": "rpm", "version": "1.2" } ], "libnetfilter_conntrack": [ { "arch": "x86_64", "epoch": null, "name": "libnetfilter_conntrack", "release": "1.el7_3", "source": "rpm", "version": "1.0.6" } ], "libnfnetlink": [ { "arch": "x86_64", "epoch": null, "name": "libnfnetlink", "release": "4.el7", "source": "rpm", "version": "1.0.1" } ], "libnfsidmap": [ { "arch": "x86_64", "epoch": null, "name": "libnfsidmap", "release": "19.el7", "source": "rpm", "version": "0.25" } ], "libnl": [ { "arch": "x86_64", "epoch": null, "name": "libnl", "release": "3.el7", "source": "rpm", "version": "1.1.4" } ], "libnl3": [ { "arch": "x86_64", "epoch": null, "name": "libnl3", "release": "4.el7", "source": "rpm", "version": "3.2.28" } ], "libnl3-cli": [ { "arch": "x86_64", "epoch": null, "name": "libnl3-cli", "release": "4.el7", "source": "rpm", "version": "3.2.28" } ], "libpath_utils": [ { "arch": "x86_64", "epoch": null, "name": "libpath_utils", "release": "32.el7", "source": "rpm", "version": "0.2.1" } ], "libpciaccess": [ { "arch": "x86_64", "epoch": null, "name": "libpciaccess", "release": "1.el7", "source": "rpm", "version": "0.14" } ], "libpipeline": [ { "arch": "x86_64", "epoch": null, "name": "libpipeline", "release": "3.el7", "source": "rpm", "version": "1.2.3" } ], "libpng": [ { "arch": "x86_64", "epoch": 2, "name": "libpng", "release": "8.el7", "source": "rpm", "version": "1.5.13" } ], "libpwquality": [ { "arch": "x86_64", "epoch": null, "name": "libpwquality", "release": "5.el7", "source": "rpm", "version": "1.2.3" } ], "libref_array": [ { "arch": "x86_64", "epoch": null, "name": "libref_array", "release": "32.el7", "source": "rpm", "version": "0.1.5" } ], "librepo": [ { "arch": "x86_64", "epoch": null, "name": "librepo", "release": "8.el7_9", "source": "rpm", "version": "1.8.1" } ], "libreport-filesystem": [ { "arch": "x86_64", "epoch": null, "name": "libreport-filesystem", "release": "53.el7.centos", "source": "rpm", "version": "2.1.11" } ], "libseccomp": [ { "arch": "x86_64", "epoch": null, "name": "libseccomp", "release": "4.el7", "source": "rpm", "version": "2.3.1" } ], "libselinux": [ { "arch": "x86_64", "epoch": null, "name": "libselinux", "release": "15.el7", "source": "rpm", "version": "2.5" } ], "libselinux-python": [ { "arch": "x86_64", "epoch": null, "name": "libselinux-python", "release": "15.el7", "source": "rpm", "version": "2.5" } ], "libselinux-utils": [ { "arch": "x86_64", "epoch": null, "name": "libselinux-utils", "release": "15.el7", "source": "rpm", "version": "2.5" } ], "libsemanage": [ { "arch": "x86_64", "epoch": null, "name": "libsemanage", "release": "14.el7", "source": "rpm", "version": "2.5" } ], "libsemanage-python": [ { "arch": "x86_64", "epoch": null, "name": "libsemanage-python", "release": "14.el7", "source": "rpm", "version": "2.5" } ], "libsepol": [ { "arch": "x86_64", "epoch": null, "name": "libsepol", "release": "10.el7", "source": "rpm", "version": "2.5" } ], "libsmartcols": [ { "arch": "x86_64", "epoch": null, "name": "libsmartcols", "release": "65.el7_9.1", "source": "rpm", "version": "2.23.2" } ], "libsolv": [ { "arch": "x86_64", "epoch": null, "name": "libsolv", "release": "4.el7", "source": "rpm", "version": "0.6.34" } ], "libss": [ { "arch": "x86_64", "epoch": null, "name": "libss", "release": "19.el7", "source": "rpm", "version": "1.42.9" } ], "libssh2": [ { "arch": "x86_64", "epoch": null, "name": "libssh2", "release": "4.el7_9.1", "source": "rpm", "version": "1.8.0" } ], "libstdc++": [ { "arch": "x86_64", "epoch": null, "name": "libstdc++", "release": "44.el7", "source": "rpm", "version": "4.8.5" } ], "libsysfs": [ { "arch": "x86_64", "epoch": null, "name": "libsysfs", "release": "16.el7", "source": "rpm", "version": "2.1.0" } ], "libtasn1": [ { "arch": "x86_64", "epoch": null, "name": "libtasn1", "release": "1.el7", "source": "rpm", "version": "4.10" } ], "libteam": [ { "arch": "x86_64", "epoch": null, "name": "libteam", "release": "3.el7", "source": "rpm", "version": "1.29" } ], "libtirpc": [ { "arch": "x86_64", "epoch": null, "name": "libtirpc", "release": "0.16.el7", "source": "rpm", "version": "0.2.4" } ], "libunistring": [ { "arch": "x86_64", "epoch": null, "name": "libunistring", "release": "9.el7", "source": "rpm", "version": "0.9.3" } ], "libuser": [ { "arch": "x86_64", "epoch": null, "name": "libuser", "release": "9.el7", "source": "rpm", "version": "0.60" } ], "libutempter": [ { "arch": "x86_64", "epoch": null, "name": "libutempter", "release": "4.el7", "source": "rpm", "version": "1.1.6" } ], "libuuid": [ { "arch": "x86_64", "epoch": null, "name": "libuuid", "release": "65.el7_9.1", "source": "rpm", "version": "2.23.2" } ], "libverto": [ { "arch": "x86_64", "epoch": null, "name": "libverto", "release": "4.el7", "source": "rpm", "version": "0.2.5" } ], "libverto-libevent": [ { "arch": "x86_64", "epoch": null, "name": "libverto-libevent", "release": "4.el7", "source": "rpm", "version": "0.2.5" } ], "libxml2": [ { "arch": "x86_64", "epoch": null, "name": "libxml2", "release": "6.el7_9.6", "source": "rpm", "version": "2.9.1" } ], "libxml2-python": [ { "arch": "x86_64", "epoch": null, "name": "libxml2-python", "release": "6.el7_9.6", "source": "rpm", "version": "2.9.1" } ], "libxslt": [ { "arch": "x86_64", "epoch": null, "name": "libxslt", "release": "6.el7", "source": "rpm", "version": "1.1.28" } ], "libyaml": [ { "arch": "x86_64", "epoch": null, "name": "libyaml", "release": "11.el7_0", "source": "rpm", "version": "0.1.4" } ], "linux-firmware": [ { "arch": "noarch", "epoch": null, "name": "linux-firmware", "release": "82.git78c0348.el7_9", "source": "rpm", "version": "20200421" } ], "logrotate": [ { "arch": "x86_64", "epoch": null, "name": "logrotate", "release": "19.el7", "source": "rpm", "version": "3.8.6" } ], "lshw": [ { "arch": "x86_64", "epoch": null, "name": "lshw", "release": "0.1.20180614git028f6b2.beaker.1.el7bkr.1", "source": "rpm", "version": "B.02.19" } ], "lsof": [ { "arch": "x86_64", "epoch": null, "name": "lsof", "release": "6.el7", "source": "rpm", "version": "4.87" } ], "lsscsi": [ { "arch": "x86_64", "epoch": null, "name": "lsscsi", "release": "6.el7", "source": "rpm", "version": "0.27" } ], "lua": [ { "arch": "x86_64", "epoch": null, "name": "lua", "release": "15.el7", "source": "rpm", "version": "5.1.4" } ], "lvm2": [ { "arch": "x86_64", "epoch": 7, "name": "lvm2", "release": "6.el7_9.5", "source": "rpm", "version": "2.02.187" } ], "lvm2-libs": [ { "arch": "x86_64", "epoch": 7, "name": "lvm2-libs", "release": "6.el7_9.5", "source": "rpm", "version": "2.02.187" } ], "lz4": [ { "arch": "x86_64", "epoch": null, "name": "lz4", "release": "1.el7", "source": "rpm", "version": "1.8.3" } ], "lzo": [ { "arch": "x86_64", "epoch": null, "name": "lzo", "release": "8.el7", "source": "rpm", "version": "2.06" } ], "mailcap": [ { "arch": "noarch", "epoch": null, "name": "mailcap", "release": "2.el7", "source": "rpm", "version": "2.1.41" } ], "make": [ { "arch": "x86_64", "epoch": 1, "name": "make", "release": "24.el7", "source": "rpm", "version": "3.82" } ], "man-db": [ { "arch": "x86_64", "epoch": null, "name": "man-db", "release": "11.el7", "source": "rpm", "version": "2.6.3" } ], "mariadb-libs": [ { "arch": "x86_64", "epoch": 1, "name": "mariadb-libs", "release": "1.el7", "source": "rpm", "version": "5.5.68" } ], "mdadm": [ { "arch": "x86_64", "epoch": null, "name": "mdadm", "release": "9.el7_9", "source": "rpm", "version": "4.1" } ], "microcode_ctl": [ { "arch": "x86_64", "epoch": 2, "name": "microcode_ctl", "release": "73.20.el7_9", "source": "rpm", "version": "2.1" } ], "mokutil": [ { "arch": "x86_64", "epoch": null, "name": "mokutil", "release": "8.el7", "source": "rpm", "version": "15" } ], "mozjs17": [ { "arch": "x86_64", "epoch": null, "name": "mozjs17", "release": "20.el7", "source": "rpm", "version": "17.0.0" } ], "mpfr": [ { "arch": "x86_64", "epoch": null, "name": "mpfr", "release": "4.el7", "source": "rpm", "version": "3.1.1" } ], "ncurses": [ { "arch": "x86_64", "epoch": null, "name": "ncurses", "release": "14.20130511.el7_4", "source": "rpm", "version": "5.9" } ], "ncurses-base": [ { "arch": "noarch", "epoch": null, "name": "ncurses-base", "release": "14.20130511.el7_4", "source": "rpm", "version": "5.9" } ], "ncurses-libs": [ { "arch": "x86_64", "epoch": null, "name": "ncurses-libs", "release": "14.20130511.el7_4", "source": "rpm", "version": "5.9" } ], "net-tools": [ { "arch": "x86_64", "epoch": null, "name": "net-tools", "release": "0.25.20131004git.el7", "source": "rpm", "version": "2.0" } ], "newt": [ { "arch": "x86_64", "epoch": null, "name": "newt", "release": "4.el7", "source": "rpm", "version": "0.52.15" } ], "newt-python": [ { "arch": "x86_64", "epoch": null, "name": "newt-python", "release": "4.el7", "source": "rpm", "version": "0.52.15" } ], "nfs-utils": [ { "arch": "x86_64", "epoch": 1, "name": "nfs-utils", "release": "0.68.el7.2", "source": "rpm", "version": "1.3.0" } ], "nspr": [ { "arch": "x86_64", "epoch": null, "name": "nspr", "release": "1.el7_9", "source": "rpm", "version": "4.35.0" } ], "nss": [ { "arch": "x86_64", "epoch": null, "name": "nss", "release": "2.el7_9", "source": "rpm", "version": "3.90.0" } ], "nss-pem": [ { "arch": "x86_64", "epoch": null, "name": "nss-pem", "release": "7.el7_9.1", "source": "rpm", "version": "1.0.3" } ], "nss-softokn": [ { "arch": "x86_64", "epoch": null, "name": "nss-softokn", "release": "6.el7_9", "source": "rpm", "version": "3.90.0" } ], "nss-softokn-freebl": [ { "arch": "x86_64", "epoch": null, "name": "nss-softokn-freebl", "release": "6.el7_9", "source": "rpm", "version": "3.90.0" } ], "nss-sysinit": [ { "arch": "x86_64", "epoch": null, "name": "nss-sysinit", "release": "2.el7_9", "source": "rpm", "version": "3.90.0" } ], "nss-tools": [ { "arch": "x86_64", "epoch": null, "name": "nss-tools", "release": "2.el7_9", "source": "rpm", "version": "3.90.0" } ], "nss-util": [ { "arch": "x86_64", "epoch": null, "name": "nss-util", "release": "1.el7_9", "source": "rpm", "version": "3.90.0" } ], "numactl-libs": [ { "arch": "x86_64", "epoch": null, "name": "numactl-libs", "release": "5.el7", "source": "rpm", "version": "2.0.12" } ], "openldap": [ { "arch": "x86_64", "epoch": null, "name": "openldap", "release": "25.el7_9", "source": "rpm", "version": "2.4.44" } ], "openssh": [ { "arch": "x86_64", "epoch": null, "name": "openssh", "release": "23.el7_9", "source": "rpm", "version": "7.4p1" } ], "openssh-clients": [ { "arch": "x86_64", "epoch": null, "name": "openssh-clients", "release": "23.el7_9", "source": "rpm", "version": "7.4p1" } ], "openssh-server": [ { "arch": "x86_64", "epoch": null, "name": "openssh-server", "release": "23.el7_9", "source": "rpm", "version": "7.4p1" } ], "openssl": [ { "arch": "x86_64", "epoch": 1, "name": "openssl", "release": "26.el7_9", "source": "rpm", "version": "1.0.2k" } ], "openssl-libs": [ { "arch": "x86_64", "epoch": 1, "name": "openssl-libs", "release": "26.el7_9", "source": "rpm", "version": "1.0.2k" } ], "os-prober": [ { "arch": "x86_64", "epoch": null, "name": "os-prober", "release": "9.el7", "source": "rpm", "version": "1.58" } ], "p11-kit": [ { "arch": "x86_64", "epoch": null, "name": "p11-kit", "release": "3.el7", "source": "rpm", "version": "0.23.5" } ], "p11-kit-trust": [ { "arch": "x86_64", "epoch": null, "name": "p11-kit-trust", "release": "3.el7", "source": "rpm", "version": "0.23.5" } ], "pam": [ { "arch": "x86_64", "epoch": null, "name": "pam", "release": "23.el7", "source": "rpm", "version": "1.1.8" } ], "parted": [ { "arch": "x86_64", "epoch": null, "name": "parted", "release": "32.el7", "source": "rpm", "version": "3.1" } ], "passwd": [ { "arch": "x86_64", "epoch": null, "name": "passwd", "release": "6.el7", "source": "rpm", "version": "0.79" } ], "pciutils-libs": [ { "arch": "x86_64", "epoch": null, "name": "pciutils-libs", "release": "3.el7", "source": "rpm", "version": "3.5.1" } ], "pcre": [ { "arch": "x86_64", "epoch": null, "name": "pcre", "release": "17.el7", "source": "rpm", "version": "8.32" } ], "perl": [ { "arch": "x86_64", "epoch": 4, "name": "perl", "release": "299.el7_9", "source": "rpm", "version": "5.16.3" } ], "perl-Business-ISBN": [ { "arch": "noarch", "epoch": null, "name": "perl-Business-ISBN", "release": "2.el7", "source": "rpm", "version": "2.06" } ], "perl-Business-ISBN-Data": [ { "arch": "noarch", "epoch": null, "name": "perl-Business-ISBN-Data", "release": "2.el7", "source": "rpm", "version": "20120719.001" } ], "perl-CGI": [ { "arch": "noarch", "epoch": null, "name": "perl-CGI", "release": "4.el7", "source": "rpm", "version": "3.63" } ], "perl-Carp": [ { "arch": "noarch", "epoch": null, "name": "perl-Carp", "release": "244.el7", "source": "rpm", "version": "1.26" } ], "perl-Compress-Raw-Bzip2": [ { "arch": "x86_64", "epoch": null, "name": "perl-Compress-Raw-Bzip2", "release": "3.el7", "source": "rpm", "version": "2.061" } ], "perl-Compress-Raw-Zlib": [ { "arch": "x86_64", "epoch": 1, "name": "perl-Compress-Raw-Zlib", "release": "4.el7", "source": "rpm", "version": "2.061" } ], "perl-Data-Dumper": [ { "arch": "x86_64", "epoch": null, "name": "perl-Data-Dumper", "release": "3.el7", "source": "rpm", "version": "2.145" } ], "perl-Date-Manip": [ { "arch": "noarch", "epoch": null, "name": "perl-Date-Manip", "release": "2.el7", "source": "rpm", "version": "6.41" } ], "perl-Digest": [ { "arch": "noarch", "epoch": null, "name": "perl-Digest", "release": "245.el7", "source": "rpm", "version": "1.17" } ], "perl-Digest-MD5": [ { "arch": "x86_64", "epoch": null, "name": "perl-Digest-MD5", "release": "3.el7", "source": "rpm", "version": "2.52" } ], "perl-Encode": [ { "arch": "x86_64", "epoch": null, "name": "perl-Encode", "release": "7.el7", "source": "rpm", "version": "2.51" } ], "perl-Encode-Locale": [ { "arch": "noarch", "epoch": null, "name": "perl-Encode-Locale", "release": "5.el7", "source": "rpm", "version": "1.03" } ], "perl-Error": [ { "arch": "noarch", "epoch": 1, "name": "perl-Error", "release": "2.el7", "source": "rpm", "version": "0.17020" } ], "perl-Exporter": [ { "arch": "noarch", "epoch": null, "name": "perl-Exporter", "release": "3.el7", "source": "rpm", "version": "5.68" } ], "perl-FCGI": [ { "arch": "x86_64", "epoch": 1, "name": "perl-FCGI", "release": "8.el7", "source": "rpm", "version": "0.74" } ], "perl-File-Listing": [ { "arch": "noarch", "epoch": null, "name": "perl-File-Listing", "release": "7.el7", "source": "rpm", "version": "6.04" } ], "perl-File-Path": [ { "arch": "noarch", "epoch": null, "name": "perl-File-Path", "release": "2.el7", "source": "rpm", "version": "2.09" } ], "perl-File-Temp": [ { "arch": "noarch", "epoch": null, "name": "perl-File-Temp", "release": "3.el7", "source": "rpm", "version": "0.23.01" } ], "perl-Filter": [ { "arch": "x86_64", "epoch": null, "name": "perl-Filter", "release": "3.el7", "source": "rpm", "version": "1.49" } ], "perl-Font-AFM": [ { "arch": "noarch", "epoch": null, "name": "perl-Font-AFM", "release": "13.el7", "source": "rpm", "version": "1.20" } ], "perl-Getopt-Long": [ { "arch": "noarch", "epoch": null, "name": "perl-Getopt-Long", "release": "3.el7", "source": "rpm", "version": "2.40" } ], "perl-Git": [ { "arch": "noarch", "epoch": null, "name": "perl-Git", "release": "25.el7_9", "source": "rpm", "version": "1.8.3.1" } ], "perl-HTML-Format": [ { "arch": "noarch", "epoch": null, "name": "perl-HTML-Format", "release": "7.el7", "source": "rpm", "version": "2.10" } ], "perl-HTML-Parser": [ { "arch": "x86_64", "epoch": null, "name": "perl-HTML-Parser", "release": "4.el7", "source": "rpm", "version": "3.71" } ], "perl-HTML-Tagset": [ { "arch": "noarch", "epoch": null, "name": "perl-HTML-Tagset", "release": "15.el7", "source": "rpm", "version": "3.20" } ], "perl-HTML-Tree": [ { "arch": "noarch", "epoch": 1, "name": "perl-HTML-Tree", "release": "2.el7", "source": "rpm", "version": "5.03" } ], "perl-HTTP-Cookies": [ { "arch": "noarch", "epoch": null, "name": "perl-HTTP-Cookies", "release": "5.el7", "source": "rpm", "version": "6.01" } ], "perl-HTTP-Daemon": [ { "arch": "noarch", "epoch": null, "name": "perl-HTTP-Daemon", "release": "8.el7", "source": "rpm", "version": "6.01" } ], "perl-HTTP-Date": [ { "arch": "noarch", "epoch": null, "name": "perl-HTTP-Date", "release": "8.el7", "source": "rpm", "version": "6.02" } ], "perl-HTTP-Message": [ { "arch": "noarch", "epoch": null, "name": "perl-HTTP-Message", "release": "6.el7", "source": "rpm", "version": "6.06" } ], "perl-HTTP-Negotiate": [ { "arch": "noarch", "epoch": null, "name": "perl-HTTP-Negotiate", "release": "5.el7", "source": "rpm", "version": "6.01" } ], "perl-HTTP-Tiny": [ { "arch": "noarch", "epoch": null, "name": "perl-HTTP-Tiny", "release": "3.el7", "source": "rpm", "version": "0.033" } ], "perl-IO-Compress": [ { "arch": "noarch", "epoch": null, "name": "perl-IO-Compress", "release": "2.el7", "source": "rpm", "version": "2.061" } ], "perl-IO-HTML": [ { "arch": "noarch", "epoch": null, "name": "perl-IO-HTML", "release": "2.el7", "source": "rpm", "version": "1.00" } ], "perl-IO-Socket-IP": [ { "arch": "noarch", "epoch": null, "name": "perl-IO-Socket-IP", "release": "5.el7", "source": "rpm", "version": "0.21" } ], "perl-IO-Socket-SSL": [ { "arch": "noarch", "epoch": null, "name": "perl-IO-Socket-SSL", "release": "7.el7", "source": "rpm", "version": "1.94" } ], "perl-IO-Zlib": [ { "arch": "noarch", "epoch": 1, "name": "perl-IO-Zlib", "release": "299.el7_9", "source": "rpm", "version": "1.10" } ], "perl-IO-stringy": [ { "arch": "noarch", "epoch": null, "name": "perl-IO-stringy", "release": "22.el7", "source": "rpm", "version": "2.110" } ], "perl-LWP-MediaTypes": [ { "arch": "noarch", "epoch": null, "name": "perl-LWP-MediaTypes", "release": "2.el7", "source": "rpm", "version": "6.02" } ], "perl-LWP-Protocol-https": [ { "arch": "noarch", "epoch": null, "name": "perl-LWP-Protocol-https", "release": "4.el7", "source": "rpm", "version": "6.04" } ], "perl-Module-Load": [ { "arch": "noarch", "epoch": 1, "name": "perl-Module-Load", "release": "3.el7", "source": "rpm", "version": "0.24" } ], "perl-Mozilla-CA": [ { "arch": "noarch", "epoch": null, "name": "perl-Mozilla-CA", "release": "5.el7", "source": "rpm", "version": "20130114" } ], "perl-Net-HTTP": [ { "arch": "noarch", "epoch": null, "name": "perl-Net-HTTP", "release": "2.el7", "source": "rpm", "version": "6.06" } ], "perl-Net-LibIDN": [ { "arch": "x86_64", "epoch": null, "name": "perl-Net-LibIDN", "release": "15.el7", "source": "rpm", "version": "0.12" } ], "perl-Net-SSLeay": [ { "arch": "x86_64", "epoch": null, "name": "perl-Net-SSLeay", "release": "6.el7", "source": "rpm", "version": "1.55" } ], "perl-PathTools": [ { "arch": "x86_64", "epoch": null, "name": "perl-PathTools", "release": "5.el7", "source": "rpm", "version": "3.40" } ], "perl-Pod-Escapes": [ { "arch": "noarch", "epoch": 1, "name": "perl-Pod-Escapes", "release": "299.el7_9", "source": "rpm", "version": "1.04" } ], "perl-Pod-Perldoc": [ { "arch": "noarch", "epoch": null, "name": "perl-Pod-Perldoc", "release": "4.el7", "source": "rpm", "version": "3.20" } ], "perl-Pod-Simple": [ { "arch": "noarch", "epoch": 1, "name": "perl-Pod-Simple", "release": "4.el7", "source": "rpm", "version": "3.28" } ], "perl-Pod-Usage": [ { "arch": "noarch", "epoch": null, "name": "perl-Pod-Usage", "release": "3.el7", "source": "rpm", "version": "1.63" } ], "perl-Scalar-List-Utils": [ { "arch": "x86_64", "epoch": null, "name": "perl-Scalar-List-Utils", "release": "248.el7", "source": "rpm", "version": "1.27" } ], "perl-Socket": [ { "arch": "x86_64", "epoch": null, "name": "perl-Socket", "release": "5.el7", "source": "rpm", "version": "2.010" } ], "perl-Storable": [ { "arch": "x86_64", "epoch": null, "name": "perl-Storable", "release": "3.el7", "source": "rpm", "version": "2.45" } ], "perl-TermReadKey": [ { "arch": "x86_64", "epoch": null, "name": "perl-TermReadKey", "release": "20.el7", "source": "rpm", "version": "2.30" } ], "perl-Text-ParseWords": [ { "arch": "noarch", "epoch": null, "name": "perl-Text-ParseWords", "release": "4.el7", "source": "rpm", "version": "3.29" } ], "perl-Time-HiRes": [ { "arch": "x86_64", "epoch": 4, "name": "perl-Time-HiRes", "release": "3.el7", "source": "rpm", "version": "1.9725" } ], "perl-Time-Local": [ { "arch": "noarch", "epoch": null, "name": "perl-Time-Local", "release": "2.el7", "source": "rpm", "version": "1.2300" } ], "perl-TimeDate": [ { "arch": "noarch", "epoch": 1, "name": "perl-TimeDate", "release": "2.el7", "source": "rpm", "version": "2.30" } ], "perl-URI": [ { "arch": "noarch", "epoch": null, "name": "perl-URI", "release": "9.el7", "source": "rpm", "version": "1.60" } ], "perl-WWW-RobotRules": [ { "arch": "noarch", "epoch": null, "name": "perl-WWW-RobotRules", "release": "5.el7", "source": "rpm", "version": "6.02" } ], "perl-XML-LibXML": [ { "arch": "x86_64", "epoch": 1, "name": "perl-XML-LibXML", "release": "5.el7", "source": "rpm", "version": "2.0018" } ], "perl-XML-NamespaceSupport": [ { "arch": "noarch", "epoch": null, "name": "perl-XML-NamespaceSupport", "release": "10.el7", "source": "rpm", "version": "1.11" } ], "perl-XML-Parser": [ { "arch": "x86_64", "epoch": null, "name": "perl-XML-Parser", "release": "10.el7", "source": "rpm", "version": "2.41" } ], "perl-XML-SAX": [ { "arch": "noarch", "epoch": null, "name": "perl-XML-SAX", "release": "9.el7", "source": "rpm", "version": "0.99" } ], "perl-XML-SAX-Base": [ { "arch": "noarch", "epoch": null, "name": "perl-XML-SAX-Base", "release": "7.el7", "source": "rpm", "version": "1.08" } ], "perl-XML-Twig": [ { "arch": "noarch", "epoch": null, "name": "perl-XML-Twig", "release": "2.el7", "source": "rpm", "version": "3.44" } ], "perl-constant": [ { "arch": "noarch", "epoch": null, "name": "perl-constant", "release": "2.el7", "source": "rpm", "version": "1.27" } ], "perl-libs": [ { "arch": "x86_64", "epoch": 4, "name": "perl-libs", "release": "299.el7_9", "source": "rpm", "version": "5.16.3" } ], "perl-libwww-perl": [ { "arch": "noarch", "epoch": null, "name": "perl-libwww-perl", "release": "2.el7", "source": "rpm", "version": "6.05" } ], "perl-macros": [ { "arch": "x86_64", "epoch": 4, "name": "perl-macros", "release": "299.el7_9", "source": "rpm", "version": "5.16.3" } ], "perl-parent": [ { "arch": "noarch", "epoch": 1, "name": "perl-parent", "release": "244.el7", "source": "rpm", "version": "0.225" } ], "perl-podlators": [ { "arch": "noarch", "epoch": null, "name": "perl-podlators", "release": "3.el7", "source": "rpm", "version": "2.5.1" } ], "perl-threads": [ { "arch": "x86_64", "epoch": null, "name": "perl-threads", "release": "4.el7", "source": "rpm", "version": "1.87" } ], "perl-threads-shared": [ { "arch": "x86_64", "epoch": null, "name": "perl-threads-shared", "release": "6.el7", "source": "rpm", "version": "1.43" } ], "pinentry": [ { "arch": "x86_64", "epoch": null, "name": "pinentry", "release": "17.el7", "source": "rpm", "version": "0.8.1" } ], "pkgconfig": [ { "arch": "x86_64", "epoch": 1, "name": "pkgconfig", "release": "4.el7", "source": "rpm", "version": "0.27.1" } ], "plymouth": [ { "arch": "x86_64", "epoch": null, "name": "plymouth", "release": "0.34.20140113.el7.centos", "source": "rpm", "version": "0.8.9" } ], "plymouth-core-libs": [ { "arch": "x86_64", "epoch": null, "name": "plymouth-core-libs", "release": "0.34.20140113.el7.centos", "source": "rpm", "version": "0.8.9" } ], "plymouth-scripts": [ { "arch": "x86_64", "epoch": null, "name": "plymouth-scripts", "release": "0.34.20140113.el7.centos", "source": "rpm", "version": "0.8.9" } ], "policycoreutils": [ { "arch": "x86_64", "epoch": null, "name": "policycoreutils", "release": "34.el7", "source": "rpm", "version": "2.5" } ], "policycoreutils-python": [ { "arch": "x86_64", "epoch": null, "name": "policycoreutils-python", "release": "34.el7", "source": "rpm", "version": "2.5" } ], "polkit": [ { "arch": "x86_64", "epoch": null, "name": "polkit", "release": "26.el7_9.1", "source": "rpm", "version": "0.112" } ], "polkit-pkla-compat": [ { "arch": "x86_64", "epoch": null, "name": "polkit-pkla-compat", "release": "4.el7", "source": "rpm", "version": "0.1" } ], "popt": [ { "arch": "x86_64", "epoch": null, "name": "popt", "release": "16.el7", "source": "rpm", "version": "1.13" } ], "postfix": [ { "arch": "x86_64", "epoch": 2, "name": "postfix", "release": "9.el7", "source": "rpm", "version": "2.10.1" } ], "procps-ng": [ { "arch": "x86_64", "epoch": null, "name": "procps-ng", "release": "28.el7", "source": "rpm", "version": "3.3.10" } ], "pth": [ { "arch": "x86_64", "epoch": null, "name": "pth", "release": "23.el7", "source": "rpm", "version": "2.0.7" } ], "pygpgme": [ { "arch": "x86_64", "epoch": null, "name": "pygpgme", "release": "9.el7", "source": "rpm", "version": "0.3" } ], "pyliblzma": [ { "arch": "x86_64", "epoch": null, "name": "pyliblzma", "release": "11.el7", "source": "rpm", "version": "0.5.3" } ], "pyparsing": [ { "arch": "noarch", "epoch": null, "name": "pyparsing", "release": "9.el7", "source": "rpm", "version": "1.5.6" } ], "pyparted": [ { "arch": "x86_64", "epoch": 1, "name": "pyparted", "release": "15.el7", "source": "rpm", "version": "3.9" } ], "pyserial": [ { "arch": "noarch", "epoch": null, "name": "pyserial", "release": "6.el7", "source": "rpm", "version": "2.6" } ], "python": [ { "arch": "x86_64", "epoch": null, "name": "python", "release": "94.el7_9", "source": "rpm", "version": "2.7.5" } ], "python-IPy": [ { "arch": "noarch", "epoch": null, "name": "python-IPy", "release": "6.el7", "source": "rpm", "version": "0.75" } ], "python-babel": [ { "arch": "noarch", "epoch": null, "name": "python-babel", "release": "8.el7", "source": "rpm", "version": "0.9.6" } ], "python-backports": [ { "arch": "x86_64", "epoch": null, "name": "python-backports", "release": "8.el7", "source": "rpm", "version": "1.0" } ], "python-backports-ssl_match_hostname": [ { "arch": "noarch", "epoch": null, "name": "python-backports-ssl_match_hostname", "release": "1.el7", "source": "rpm", "version": "3.5.0.1" } ], "python-chardet": [ { "arch": "noarch", "epoch": null, "name": "python-chardet", "release": "3.el7", "source": "rpm", "version": "2.2.1" } ], "python-configobj": [ { "arch": "noarch", "epoch": null, "name": "python-configobj", "release": "7.el7", "source": "rpm", "version": "4.7.2" } ], "python-configshell": [ { "arch": "noarch", "epoch": 1, "name": "python-configshell", "release": "1.el7", "source": "rpm", "version": "1.1.26" } ], "python-decorator": [ { "arch": "noarch", "epoch": null, "name": "python-decorator", "release": "3.el7", "source": "rpm", "version": "3.4.0" } ], "python-enum34": [ { "arch": "noarch", "epoch": null, "name": "python-enum34", "release": "1.el7", "source": "rpm", "version": "1.0.4" } ], "python-ethtool": [ { "arch": "x86_64", "epoch": null, "name": "python-ethtool", "release": "8.el7", "source": "rpm", "version": "0.8" } ], "python-firewall": [ { "arch": "noarch", "epoch": null, "name": "python-firewall", "release": "13.el7_9", "source": "rpm", "version": "0.6.3" } ], "python-gobject-base": [ { "arch": "x86_64", "epoch": null, "name": "python-gobject-base", "release": "1.el7_4.1", "source": "rpm", "version": "3.22.0" } ], "python-iniparse": [ { "arch": "noarch", "epoch": null, "name": "python-iniparse", "release": "9.el7", "source": "rpm", "version": "0.4" } ], "python-ipaddress": [ { "arch": "noarch", "epoch": null, "name": "python-ipaddress", "release": "2.el7", "source": "rpm", "version": "1.0.16" } ], "python-jinja2": [ { "arch": "noarch", "epoch": null, "name": "python-jinja2", "release": "4.el7", "source": "rpm", "version": "2.7.2" } ], "python-jsonpatch": [ { "arch": "noarch", "epoch": null, "name": "python-jsonpatch", "release": "4.el7", "source": "rpm", "version": "1.2" } ], "python-jsonpointer": [ { "arch": "noarch", "epoch": null, "name": "python-jsonpointer", "release": "2.el7", "source": "rpm", "version": "1.9" } ], "python-kitchen": [ { "arch": "noarch", "epoch": null, "name": "python-kitchen", "release": "5.el7", "source": "rpm", "version": "1.1.1" } ], "python-kmod": [ { "arch": "x86_64", "epoch": null, "name": "python-kmod", "release": "4.el7", "source": "rpm", "version": "0.9" } ], "python-libs": [ { "arch": "x86_64", "epoch": null, "name": "python-libs", "release": "94.el7_9", "source": "rpm", "version": "2.7.5" } ], "python-linux-procfs": [ { "arch": "noarch", "epoch": null, "name": "python-linux-procfs", "release": "4.el7", "source": "rpm", "version": "0.4.11" } ], "python-lxml": [ { "arch": "x86_64", "epoch": null, "name": "python-lxml", "release": "4.el7", "source": "rpm", "version": "3.2.1" } ], "python-markupsafe": [ { "arch": "x86_64", "epoch": null, "name": "python-markupsafe", "release": "10.el7", "source": "rpm", "version": "0.11" } ], "python-perf": [ { "arch": "x86_64", "epoch": null, "name": "python-perf", "release": "1160.119.1.el7", "source": "rpm", "version": "3.10.0" } ], "python-prettytable": [ { "arch": "noarch", "epoch": null, "name": "python-prettytable", "release": "3.el7", "source": "rpm", "version": "0.7.2" } ], "python-pycurl": [ { "arch": "x86_64", "epoch": null, "name": "python-pycurl", "release": "19.el7", "source": "rpm", "version": "7.19.0" } ], "python-pyudev": [ { "arch": "noarch", "epoch": null, "name": "python-pyudev", "release": "9.el7", "source": "rpm", "version": "0.15" } ], "python-requests": [ { "arch": "noarch", "epoch": null, "name": "python-requests", "release": "10.el7", "source": "rpm", "version": "2.6.0" } ], "python-rtslib": [ { "arch": "noarch", "epoch": null, "name": "python-rtslib", "release": "1.el7_9", "source": "rpm", "version": "2.1.74" } ], "python-schedutils": [ { "arch": "x86_64", "epoch": null, "name": "python-schedutils", "release": "6.el7", "source": "rpm", "version": "0.4" } ], "python-setuptools": [ { "arch": "noarch", "epoch": null, "name": "python-setuptools", "release": "7.el7", "source": "rpm", "version": "0.9.8" } ], "python-six": [ { "arch": "noarch", "epoch": null, "name": "python-six", "release": "2.el7", "source": "rpm", "version": "1.9.0" } ], "python-slip": [ { "arch": "noarch", "epoch": null, "name": "python-slip", "release": "4.el7", "source": "rpm", "version": "0.4.0" } ], "python-slip-dbus": [ { "arch": "noarch", "epoch": null, "name": "python-slip-dbus", "release": "4.el7", "source": "rpm", "version": "0.4.0" } ], "python-urlgrabber": [ { "arch": "noarch", "epoch": null, "name": "python-urlgrabber", "release": "10.el7", "source": "rpm", "version": "3.10" } ], "python-urllib3": [ { "arch": "noarch", "epoch": null, "name": "python-urllib3", "release": "7.el7", "source": "rpm", "version": "1.10.2" } ], "python-urwid": [ { "arch": "x86_64", "epoch": null, "name": "python-urwid", "release": "3.el7", "source": "rpm", "version": "1.1.1" } ], "python2-blivet3": [ { "arch": "noarch", "epoch": 1, "name": "python2-blivet3", "release": "3.el7", "source": "rpm", "version": "3.1.3" } ], "python2-blockdev": [ { "arch": "x86_64", "epoch": null, "name": "python2-blockdev", "release": "5.el7", "source": "rpm", "version": "2.18" } ], "python2-bytesize": [ { "arch": "x86_64", "epoch": null, "name": "python2-bytesize", "release": "1.el7", "source": "rpm", "version": "1.2" } ], "python2-hawkey": [ { "arch": "x86_64", "epoch": null, "name": "python2-hawkey", "release": "2.el7_9", "source": "rpm", "version": "0.22.5" } ], "python2-libdnf": [ { "arch": "x86_64", "epoch": null, "name": "python2-libdnf", "release": "2.el7_9", "source": "rpm", "version": "0.22.5" } ], "pyxattr": [ { "arch": "x86_64", "epoch": null, "name": "pyxattr", "release": "5.el7", "source": "rpm", "version": "0.5.1" } ], "qa-tools": [ { "arch": "noarch", "epoch": null, "name": "qa-tools", "release": "4.el7_9", "source": "rpm", "version": "4.1" } ], "qemu-guest-agent": [ { "arch": "x86_64", "epoch": 10, "name": "qemu-guest-agent", "release": "3.el7", "source": "rpm", "version": "2.12.0" } ], "qrencode-libs": [ { "arch": "x86_64", "epoch": null, "name": "qrencode-libs", "release": "3.el7", "source": "rpm", "version": "3.4.1" } ], "quota": [ { "arch": "x86_64", "epoch": 1, "name": "quota", "release": "19.el7", "source": "rpm", "version": "4.01" } ], "quota-nls": [ { "arch": "noarch", "epoch": 1, "name": "quota-nls", "release": "19.el7", "source": "rpm", "version": "4.01" } ], "readline": [ { "arch": "x86_64", "epoch": null, "name": "readline", "release": "11.el7", "source": "rpm", "version": "6.2" } ], "restraint": [ { "arch": "x86_64", "epoch": null, "name": "restraint", "release": "1.el7bkr", "source": "rpm", "version": "0.4.4" } ], "restraint-rhts": [ { "arch": "x86_64", "epoch": null, "name": "restraint-rhts", "release": "1.el7bkr", "source": "rpm", "version": "0.4.4" } ], "rng-tools": [ { "arch": "x86_64", "epoch": null, "name": "rng-tools", "release": "5.el7", "source": "rpm", "version": "6.3.1" } ], "rootfiles": [ { "arch": "noarch", "epoch": null, "name": "rootfiles", "release": "11.el7", "source": "rpm", "version": "8.1" } ], "rpcbind": [ { "arch": "x86_64", "epoch": null, "name": "rpcbind", "release": "49.el7", "source": "rpm", "version": "0.2.0" } ], "rpm": [ { "arch": "x86_64", "epoch": null, "name": "rpm", "release": "48.el7_9", "source": "rpm", "version": "4.11.3" } ], "rpm-build-libs": [ { "arch": "x86_64", "epoch": null, "name": "rpm-build-libs", "release": "48.el7_9", "source": "rpm", "version": "4.11.3" } ], "rpm-libs": [ { "arch": "x86_64", "epoch": null, "name": "rpm-libs", "release": "48.el7_9", "source": "rpm", "version": "4.11.3" } ], "rpm-python": [ { "arch": "x86_64", "epoch": null, "name": "rpm-python", "release": "48.el7_9", "source": "rpm", "version": "4.11.3" } ], "rsync": [ { "arch": "x86_64", "epoch": null, "name": "rsync", "release": "12.el7_9", "source": "rpm", "version": "3.1.2" } ], "rsyslog": [ { "arch": "x86_64", "epoch": null, "name": "rsyslog", "release": "57.el7_9.3", "source": "rpm", "version": "8.24.0" } ], "ruby": [ { "arch": "x86_64", "epoch": null, "name": "ruby", "release": "39.el7_9", "source": "rpm", "version": "2.0.0.648" } ], "ruby-irb": [ { "arch": "noarch", "epoch": null, "name": "ruby-irb", "release": "39.el7_9", "source": "rpm", "version": "2.0.0.648" } ], "ruby-libs": [ { "arch": "x86_64", "epoch": null, "name": "ruby-libs", "release": "39.el7_9", "source": "rpm", "version": "2.0.0.648" } ], "rubygem-bigdecimal": [ { "arch": "x86_64", "epoch": null, "name": "rubygem-bigdecimal", "release": "39.el7_9", "source": "rpm", "version": "1.2.0" } ], "rubygem-io-console": [ { "arch": "x86_64", "epoch": null, "name": "rubygem-io-console", "release": "39.el7_9", "source": "rpm", "version": "0.4.2" } ], "rubygem-json": [ { "arch": "x86_64", "epoch": null, "name": "rubygem-json", "release": "39.el7_9", "source": "rpm", "version": "1.7.7" } ], "rubygem-psych": [ { "arch": "x86_64", "epoch": null, "name": "rubygem-psych", "release": "39.el7_9", "source": "rpm", "version": "2.0.0" } ], "rubygem-rdoc": [ { "arch": "noarch", "epoch": null, "name": "rubygem-rdoc", "release": "39.el7_9", "source": "rpm", "version": "4.0.0" } ], "rubygems": [ { "arch": "noarch", "epoch": null, "name": "rubygems", "release": "39.el7_9", "source": "rpm", "version": "2.0.14.1" } ], "sed": [ { "arch": "x86_64", "epoch": null, "name": "sed", "release": "7.el7", "source": "rpm", "version": "4.2.2" } ], "selinux-policy": [ { "arch": "noarch", "epoch": null, "name": "selinux-policy", "release": "268.el7_9.2", "source": "rpm", "version": "3.13.1" } ], "selinux-policy-targeted": [ { "arch": "noarch", "epoch": null, "name": "selinux-policy-targeted", "release": "268.el7_9.2", "source": "rpm", "version": "3.13.1" } ], "setools-libs": [ { "arch": "x86_64", "epoch": null, "name": "setools-libs", "release": "4.el7", "source": "rpm", "version": "3.3.8" } ], "setup": [ { "arch": "noarch", "epoch": null, "name": "setup", "release": "11.el7", "source": "rpm", "version": "2.8.71" } ], "sg3_utils": [ { "arch": "x86_64", "epoch": 1, "name": "sg3_utils", "release": "19.el7", "source": "rpm", "version": "1.37" } ], "sg3_utils-libs": [ { "arch": "x86_64", "epoch": 1, "name": "sg3_utils-libs", "release": "19.el7", "source": "rpm", "version": "1.37" } ], "sgpio": [ { "arch": "x86_64", "epoch": null, "name": "sgpio", "release": "13.el7", "source": "rpm", "version": "1.2.0.10" } ], "shadow-utils": [ { "arch": "x86_64", "epoch": 2, "name": "shadow-utils", "release": "5.el7", "source": "rpm", "version": "4.6" } ], "shared-mime-info": [ { "arch": "x86_64", "epoch": null, "name": "shared-mime-info", "release": "5.el7", "source": "rpm", "version": "1.8" } ], "slang": [ { "arch": "x86_64", "epoch": null, "name": "slang", "release": "11.el7", "source": "rpm", "version": "2.2.4" } ], "snappy": [ { "arch": "x86_64", "epoch": null, "name": "snappy", "release": "3.el7", "source": "rpm", "version": "1.1.0" } ], "sqlite": [ { "arch": "x86_64", "epoch": null, "name": "sqlite", "release": "8.el7_7.1", "source": "rpm", "version": "3.7.17" } ], "strace": [ { "arch": "x86_64", "epoch": null, "name": "strace", "release": "7.el7_9", "source": "rpm", "version": "4.24" } ], "sudo": [ { "arch": "x86_64", "epoch": null, "name": "sudo", "release": "10.el7_9.3", "source": "rpm", "version": "1.8.23" } ], "systemd": [ { "arch": "x86_64", "epoch": null, "name": "systemd", "release": "78.el7_9.9", "source": "rpm", "version": "219" } ], "systemd-libs": [ { "arch": "x86_64", "epoch": null, "name": "systemd-libs", "release": "78.el7_9.9", "source": "rpm", "version": "219" } ], "systemd-sysv": [ { "arch": "x86_64", "epoch": null, "name": "systemd-sysv", "release": "78.el7_9.9", "source": "rpm", "version": "219" } ], "systemtap": [ { "arch": "x86_64", "epoch": null, "name": "systemtap", "release": "13.el7", "source": "rpm", "version": "4.0" } ], "systemtap-client": [ { "arch": "x86_64", "epoch": null, "name": "systemtap-client", "release": "13.el7", "source": "rpm", "version": "4.0" } ], "systemtap-devel": [ { "arch": "x86_64", "epoch": null, "name": "systemtap-devel", "release": "13.el7", "source": "rpm", "version": "4.0" } ], "systemtap-runtime": [ { "arch": "x86_64", "epoch": null, "name": "systemtap-runtime", "release": "13.el7", "source": "rpm", "version": "4.0" } ], "sysvinit-tools": [ { "arch": "x86_64", "epoch": null, "name": "sysvinit-tools", "release": "14.dsf.el7", "source": "rpm", "version": "2.88" } ], "tar": [ { "arch": "x86_64", "epoch": 2, "name": "tar", "release": "35.el7", "source": "rpm", "version": "1.26" } ], "targetcli": [ { "arch": "noarch", "epoch": null, "name": "targetcli", "release": "1.el7_9", "source": "rpm", "version": "2.1.53" } ], "tcp_wrappers": [ { "arch": "x86_64", "epoch": null, "name": "tcp_wrappers", "release": "77.el7", "source": "rpm", "version": "7.6" } ], "tcp_wrappers-libs": [ { "arch": "x86_64", "epoch": null, "name": "tcp_wrappers-libs", "release": "77.el7", "source": "rpm", "version": "7.6" } ], "teamd": [ { "arch": "x86_64", "epoch": null, "name": "teamd", "release": "3.el7", "source": "rpm", "version": "1.29" } ], "time": [ { "arch": "x86_64", "epoch": null, "name": "time", "release": "45.el7", "source": "rpm", "version": "1.7" } ], "tps-devel": [ { "arch": "noarch", "epoch": null, "name": "tps-devel", "release": "1", "source": "rpm", "version": "2.44.50" } ], "tuned": [ { "arch": "noarch", "epoch": null, "name": "tuned", "release": "12.el7_9", "source": "rpm", "version": "2.11.0" } ], "tzdata": [ { "arch": "noarch", "epoch": null, "name": "tzdata", "release": "1.el7", "source": "rpm", "version": "2024a" } ], "unzip": [ { "arch": "x86_64", "epoch": null, "name": "unzip", "release": "24.el7_9", "source": "rpm", "version": "6.0" } ], "ustr": [ { "arch": "x86_64", "epoch": null, "name": "ustr", "release": "16.el7", "source": "rpm", "version": "1.0.4" } ], "util-linux": [ { "arch": "x86_64", "epoch": null, "name": "util-linux", "release": "65.el7_9.1", "source": "rpm", "version": "2.23.2" } ], "vim-common": [ { "arch": "x86_64", "epoch": 2, "name": "vim-common", "release": "8.el7_9", "source": "rpm", "version": "7.4.629" } ], "vim-enhanced": [ { "arch": "x86_64", "epoch": 2, "name": "vim-enhanced", "release": "8.el7_9", "source": "rpm", "version": "7.4.629" } ], "vim-filesystem": [ { "arch": "x86_64", "epoch": 2, "name": "vim-filesystem", "release": "8.el7_9", "source": "rpm", "version": "7.4.629" } ], "vim-minimal": [ { "arch": "x86_64", "epoch": 2, "name": "vim-minimal", "release": "8.el7_9", "source": "rpm", "version": "7.4.629" } ], "virt-what": [ { "arch": "x86_64", "epoch": null, "name": "virt-what", "release": "4.el7_9.1", "source": "rpm", "version": "1.18" } ], "volume_key-libs": [ { "arch": "x86_64", "epoch": null, "name": "volume_key-libs", "release": "9.el7", "source": "rpm", "version": "0.3.9" } ], "wget": [ { "arch": "x86_64", "epoch": null, "name": "wget", "release": "18.el7_6.1", "source": "rpm", "version": "1.14" } ], "which": [ { "arch": "x86_64", "epoch": null, "name": "which", "release": "7.el7", "source": "rpm", "version": "2.20" } ], "wpa_supplicant": [ { "arch": "x86_64", "epoch": 1, "name": "wpa_supplicant", "release": "12.el7_9.2", "source": "rpm", "version": "2.6" } ], "xfsprogs": [ { "arch": "x86_64", "epoch": null, "name": "xfsprogs", "release": "22.el7", "source": "rpm", "version": "4.5.0" } ], "xz": [ { "arch": "x86_64", "epoch": null, "name": "xz", "release": "2.el7_9", "source": "rpm", "version": "5.2.2" } ], "xz-libs": [ { "arch": "x86_64", "epoch": null, "name": "xz-libs", "release": "2.el7_9", "source": "rpm", "version": "5.2.2" } ], "yum": [ { "arch": "noarch", "epoch": null, "name": "yum", "release": "168.el7.centos", "source": "rpm", "version": "3.4.3" } ], "yum-metadata-parser": [ { "arch": "x86_64", "epoch": null, "name": "yum-metadata-parser", "release": "10.el7", "source": "rpm", "version": "1.1.4" } ], "yum-plugin-fastestmirror": [ { "arch": "noarch", "epoch": null, "name": "yum-plugin-fastestmirror", "release": "54.el7_8", "source": "rpm", "version": "1.1.31" } ], "yum-utils": [ { "arch": "noarch", "epoch": null, "name": "yum-utils", "release": "54.el7_8", "source": "rpm", "version": "1.1.31" } ], "zip": [ { "arch": "x86_64", "epoch": null, "name": "zip", "release": "11.el7", "source": "rpm", "version": "3.0" } ], "zlib": [ { "arch": "x86_64", "epoch": null, "name": "zlib", "release": "21.el7_9", "source": "rpm", "version": "1.2.7" } ] } }, "changed": false } TASK [Set blivet package name] ************************************************* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:321 Saturday 17 August 2024 19:53:33 -0400 (0:00:01.646) 0:04:44.695 ******* ok: [managed_node2] => { "ansible_facts": { "blivet_pkg_name": [ "python2-blivet3" ] }, "changed": false } TASK [Set blivet package version] ********************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:325 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.106) 0:04:44.802 ******* ok: [managed_node2] => { "ansible_facts": { "blivet_pkg_version": "3.1.3-3.el7" }, "changed": false } TASK [Set distribution version] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:329 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.106) 0:04:44.909 ******* ok: [managed_node2] => { "ansible_facts": { "is_fedora": false, "is_rhel10": false, "is_rhel8": false, "is_rhel9": false }, "changed": false } TASK [Create one LVM logical volume under one volume group with size 5g] ******* task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:341 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.150) 0:04:45.060 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:357 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.078) 0:04:45.139 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Change volume_size to 9g] ************************************************ task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:360 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.144) 0:04:45.284 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:374 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.079) 0:04:45.363 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Test for correct handling of offline resize in safe mode] **************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:377 Saturday 17 August 2024 19:53:33 -0400 (0:00:00.080) 0:04:45.443 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up] **************************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:394 Saturday 17 August 2024 19:53:34 -0400 (0:00:00.094) 0:04:45.538 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify role results] ***************************************************** task path: /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:407 Saturday 17 August 2024 19:53:34 -0400 (0:00:00.077) 0:04:45.615 ******* skipping: [managed_node2] => { "changed": false, "skip_reason": "Conditional result was False" } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* managed_node2 : ok=1339 changed=26 unreachable=0 failed=3 skipped=1498 rescued=3 ignored=0 Saturday 17 August 2024 19:53:34 -0400 (0:00:00.022) 0:04:45.638 ******* =============================================================================== fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 7.35s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.83s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.26s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.24s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.20s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 5.11s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.66s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.52s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.52s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.49s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.48s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.22s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.14s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.14s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 4.10s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Manage the pools and volumes to match the specified state --- 3.72s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:69 fedora.linux_system_roles.storage : Update facts ------------------------ 1.77s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 fedora.linux_system_roles.storage : Update facts ------------------------ 1.69s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:222 Gather package facts ---------------------------------------------------- 1.65s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/tests/storage/tests_resize.yml:318 fedora.linux_system_roles.storage : Make sure blivet is available ------- 1.36s /tmp/collections-RCe/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2