ansible-playbook [core 2.17.13] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-lSO executable location = /usr/local/bin/ansible-playbook python version = 3.12.11 (main, Aug 14 2025, 00:00:00) [GCC 11.5.0 20240719 (Red Hat 11.5.0-11)] (/usr/bin/python3.12) jinja version = 3.1.6 libyaml = True No config file found; using defaults running playbook inside collection fedora.linux_system_roles Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'jsonl', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_default.yml **************************************************** 1 plays in /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/tests/hpc/tests_default.yml PLAY [Ensure that the role runs with default parameters] *********************** TASK [fedora.linux_system_roles.hpc : Set platform/version specific variables] *** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:3 Monday 01 September 2025 08:25:30 -0400 (0:00:00.022) 0:00:00.022 ****** included: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml for managed-node1 TASK [fedora.linux_system_roles.hpc : Ensure ansible_facts used by role] ******* task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:2 Monday 01 September 2025 08:25:30 -0400 (0:00:00.028) 0:00:00.051 ****** [WARNING]: Platform linux on host managed-node1 is using the discovered Python interpreter at /usr/bin/python3.9, but future installation of another Python interpreter could change the meaning of that path. See https://docs.ansible.com/ansible- core/2.17/reference_appendices/interpreter_discovery.html for more information. ok: [managed-node1] TASK [fedora.linux_system_roles.hpc : Check if system is ostree] *************** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:10 Monday 01 September 2025 08:25:31 -0400 (0:00:00.764) 0:00:00.816 ****** ok: [managed-node1] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.hpc : Set flag to indicate system is ostree] *** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:15 Monday 01 September 2025 08:25:31 -0400 (0:00:00.435) 0:00:01.251 ****** ok: [managed-node1] => { "ansible_facts": { "__hpc_is_ostree": false }, "changed": false } TASK [fedora.linux_system_roles.hpc : Set platform/version specific variables] *** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:19 Monday 01 September 2025 08:25:31 -0400 (0:00:00.031) 0:00:01.283 ****** skipping: [managed-node1] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [managed-node1] => (item=CentOS.yml) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__vars_file is file", "item": "CentOS.yml", "skip_reason": "Conditional result was False" } ok: [managed-node1] => (item=CentOS_9.yml) => { "ansible_facts": { "__template_packages": [], "__template_services": [] }, "ansible_included_var_files": [ "/tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } ok: [managed-node1] => (item=CentOS_9.yml) => { "ansible_facts": { "__template_packages": [], "__template_services": [] }, "ansible_included_var_files": [ "/tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/vars/CentOS_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "CentOS_9.yml" } TASK [fedora.linux_system_roles.hpc : Deploy the GPG key for RHEL EPEL repository] *** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:6 Monday 01 September 2025 08:25:31 -0400 (0:00:00.041) 0:00:01.325 ****** ok: [managed-node1] => { "changed": false } TASK [fedora.linux_system_roles.hpc : Install EPEL release package] ************ task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:12 Monday 01 September 2025 08:25:32 -0400 (0:00:00.799) 0:00:02.125 ****** ok: [managed-node1] => { "changed": false, "rc": 0, "results": [ "Installed /root/.ansible/tmp/ansible-tmp-1756729532.4172022-9241-57195450884605/epel-release-latest-9.noarchgh2wvjxz.rpm" ] } MSG: Nothing to do TASK [fedora.linux_system_roles.hpc : Deploy the GPG key for NVIDIA repositories] *** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:18 Monday 01 September 2025 08:25:34 -0400 (0:00:02.223) 0:00:04.348 ****** changed: [managed-node1] => { "changed": true } TASK [fedora.linux_system_roles.hpc : Configure the NVIDIA CUDA repository] **** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:23 Monday 01 September 2025 08:25:35 -0400 (0:00:00.487) 0:00:04.836 ****** redirecting (type: action) ansible.builtin.yum to ansible.builtin.dnf changed: [managed-node1] => { "changed": true, "repo": "nvidia-cuda", "state": "present" } TASK [fedora.linux_system_roles.hpc : Install lvm2 to get lvs command] ********* task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:33 Monday 01 September 2025 08:25:35 -0400 (0:00:00.435) 0:00:05.271 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.hpc : Get current LV size of rootlv] *********** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:39 Monday 01 September 2025 08:25:35 -0400 (0:00:00.014) 0:00:05.285 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.hpc : Get current LV size of usrlv] ************ task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:46 Monday 01 September 2025 08:25:35 -0400 (0:00:00.011) 0:00:05.297 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [Configure storage] ******************************************************* task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:53 Monday 01 September 2025 08:25:35 -0400 (0:00:00.010) 0:00:05.307 ****** skipping: [managed-node1] => { "changed": false, "false_condition": "hpc_manage_storage", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.hpc : Update RHUI packages from Microsoft repositories] *** task path: /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:81 Monday 01 September 2025 08:25:35 -0400 (0:00:00.010) 0:00:05.318 ****** fatal: [managed-node1]: FAILED! => {} MSG: The conditional check 'ansible_system_vendor == "Microsoft Corporation"' failed. The error was: error while evaluating conditional (ansible_system_vendor == "Microsoft Corporation"): 'ansible_system_vendor' is undefined The error appears to be in '/tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml': line 81, column 3, but may be elsewhere in the file depending on the exact syntax problem. The offending line appears to be: # This is only needed for MicrosoftAzure VMs - name: Update RHUI packages from Microsoft repositories ^ here PLAY RECAP ********************************************************************* managed-node1 : ok=9 changed=2 unreachable=0 failed=1 skipped=4 rescued=0 ignored=0 SYSTEM ROLES ERRORS BEGIN v1 [ { "ansible_version": "2.17.13", "end_time": "2025-09-01T12:25:35.558044+00:00Z", "host": "managed-node1", "message": "The conditional check 'ansible_system_vendor == \"Microsoft Corporation\"' failed. The error was: error while evaluating conditional (ansible_system_vendor == \"Microsoft Corporation\"): 'ansible_system_vendor' is undefined\n\nThe error appears to be in '/tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml': line 81, column 3, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n# This is only needed for MicrosoftAzure VMs\n- name: Update RHUI packages from Microsoft repositories\n ^ here\n", "start_time": "2025-09-01T12:25:35.549499+00:00Z", "task_name": "Update RHUI packages from Microsoft repositories", "task_path": "/tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:81" } ] SYSTEM ROLES ERRORS END v1 TASKS RECAP ******************************************************************** Monday 01 September 2025 08:25:35 -0400 (0:00:00.009) 0:00:05.328 ****** =============================================================================== fedora.linux_system_roles.hpc : Install EPEL release package ------------ 2.22s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:12 fedora.linux_system_roles.hpc : Deploy the GPG key for RHEL EPEL repository --- 0.80s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:6 fedora.linux_system_roles.hpc : Ensure ansible_facts used by role ------- 0.76s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:2 fedora.linux_system_roles.hpc : Deploy the GPG key for NVIDIA repositories --- 0.49s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:18 fedora.linux_system_roles.hpc : Check if system is ostree --------------- 0.44s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:10 fedora.linux_system_roles.hpc : Configure the NVIDIA CUDA repository ---- 0.44s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:23 fedora.linux_system_roles.hpc : Set platform/version specific variables --- 0.04s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:19 fedora.linux_system_roles.hpc : Set flag to indicate system is ostree --- 0.03s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/set_vars.yml:15 fedora.linux_system_roles.hpc : Set platform/version specific variables --- 0.03s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:3 fedora.linux_system_roles.hpc : Install lvm2 to get lvs command --------- 0.01s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:33 fedora.linux_system_roles.hpc : Get current LV size of rootlv ----------- 0.01s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:39 fedora.linux_system_roles.hpc : Get current LV size of usrlv ------------ 0.01s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:46 Configure storage ------------------------------------------------------- 0.01s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:53 fedora.linux_system_roles.hpc : Update RHUI packages from Microsoft repositories --- 0.01s /tmp/collections-lSO/ansible_collections/fedora/linux_system_roles/roles/hpc/tasks/main.yml:81 Sep 01 08:25:29 managed-node1 python3.9[8810]: ansible-ansible.legacy.setup Invoked with gather_subset=['all'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Sep 01 08:25:29 managed-node1 sshd[8863]: Accepted publickey for root from 10.31.9.95 port 56060 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Sep 01 08:25:29 managed-node1 systemd-logind[606]: New session 13 of user root. ░░ Subject: A new session 13 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 13 has been created for the user root. ░░ ░░ The leading process of the session is 8863. Sep 01 08:25:29 managed-node1 systemd[1]: Started Session 13 of User root. ░░ Subject: A start job for unit session-13.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-13.scope has finished successfully. ░░ ░░ The job identifier is 1851. Sep 01 08:25:29 managed-node1 sshd[8863]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Sep 01 08:25:29 managed-node1 sshd[8866]: Received disconnect from 10.31.9.95 port 56060:11: disconnected by user Sep 01 08:25:29 managed-node1 sshd[8866]: Disconnected from user root 10.31.9.95 port 56060 Sep 01 08:25:29 managed-node1 sshd[8863]: pam_unix(sshd:session): session closed for user root Sep 01 08:25:29 managed-node1 systemd[1]: session-13.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-13.scope has successfully entered the 'dead' state. Sep 01 08:25:29 managed-node1 systemd-logind[606]: Session 13 logged out. Waiting for processes to exit. Sep 01 08:25:29 managed-node1 systemd-logind[606]: Removed session 13. ░░ Subject: Session 13 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 13 has been terminated. Sep 01 08:25:30 managed-node1 python3.9[9064]: ansible-setup Invoked with gather_subset=['!all', '!min', 'distribution', 'distribution_major_version', 'distribution_version', 'os_family'] gather_timeout=10 filter=[] fact_path=/etc/ansible/facts.d Sep 01 08:25:31 managed-node1 python3.9[9215]: ansible-stat Invoked with path=/run/ostree-booted follow=False get_checksum=True get_mime=True get_attributes=True checksum_algorithm=sha1 Sep 01 08:25:31 managed-node1 python3.9[9364]: ansible-rpm_key Invoked with key=https://dl.fedoraproject.org/pub/epel/RPM-GPG-KEY-EPEL-9 state=present validate_certs=True fingerprint=None Sep 01 08:25:32 managed-node1 python3.9[9518]: ansible-ansible.legacy.setup Invoked with filter=['ansible_pkg_mgr'] gather_subset=['!all'] gather_timeout=10 fact_path=/etc/ansible/facts.d Sep 01 08:25:33 managed-node1 python3.9[9595]: ansible-ansible.legacy.dnf Invoked with name=['https://dl.fedoraproject.org/pub/epel/epel-release-latest-9.noarch.rpm'] state=present allow_downgrade=False allowerasing=False autoremove=False bugfix=False cacheonly=False disable_gpg_check=False disable_plugin=[] disablerepo=[] download_only=False enable_plugin=[] enablerepo=[] exclude=[] installroot=/ install_repoquery=True install_weak_deps=True security=False skip_broken=False update_cache=False update_only=False validate_certs=True sslverify=True lock_timeout=30 use_backend=auto best=None conf_file=None disable_excludes=None download_dir=None list=None nobest=None releasever=None Sep 01 08:25:34 managed-node1 python3.9[9745]: ansible-rpm_key Invoked with key=https://developer.download.nvidia.com/compute/cuda/repos/rhel9/x86_64/D42D0685.pub state=present validate_certs=True fingerprint=None Sep 01 08:25:35 managed-node1 python3.9[9900]: ansible-yum_repository Invoked with name=nvidia-cuda description=NVIDIA CUDA repository baseurl=['https://developer.download.nvidia.com/compute/cuda/repos/rhel9/x86_64'] gpgcheck=True reposdir=/etc/yum.repos.d state=present unsafe_writes=False bandwidth=None cost=None deltarpm_metadata_percentage=None deltarpm_percentage=None enabled=None enablegroups=None exclude=None failovermethod=None file=None gpgcakey=None gpgkey=None module_hotfixes=None http_caching=None include=None includepkgs=None ip_resolve=None keepalive=None keepcache=None metadata_expire=None metadata_expire_filter=None metalink=None mirrorlist=None mirrorlist_expire=None password=NOT_LOGGING_PARAMETER priority=None protect=None proxy=None proxy_password=NOT_LOGGING_PARAMETER proxy_username=None repo_gpgcheck=None retries=None s3_enabled=None skip_if_unavailable=None sslcacert=None ssl_check_cert_permissions=None sslclientcert=None sslclientkey=None sslverify=None throttle=None timeout=None ui_repoid_vars=None username=None async=None mode=None owner=None group=None seuser=None serole=None selevel=None setype=None attributes=None Sep 01 08:25:35 managed-node1 sshd[9925]: Accepted publickey for root from 10.31.9.95 port 56062 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Sep 01 08:25:35 managed-node1 systemd-logind[606]: New session 14 of user root. ░░ Subject: A new session 14 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 14 has been created for the user root. ░░ ░░ The leading process of the session is 9925. Sep 01 08:25:35 managed-node1 systemd[1]: Started Session 14 of User root. ░░ Subject: A start job for unit session-14.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-14.scope has finished successfully. ░░ ░░ The job identifier is 1920. Sep 01 08:25:35 managed-node1 sshd[9925]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0) Sep 01 08:25:35 managed-node1 sshd[9928]: Received disconnect from 10.31.9.95 port 56062:11: disconnected by user Sep 01 08:25:35 managed-node1 sshd[9928]: Disconnected from user root 10.31.9.95 port 56062 Sep 01 08:25:35 managed-node1 sshd[9925]: pam_unix(sshd:session): session closed for user root Sep 01 08:25:35 managed-node1 systemd-logind[606]: Session 14 logged out. Waiting for processes to exit. Sep 01 08:25:35 managed-node1 systemd[1]: session-14.scope: Deactivated successfully. ░░ Subject: Unit succeeded ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ The unit session-14.scope has successfully entered the 'dead' state. Sep 01 08:25:35 managed-node1 systemd-logind[606]: Removed session 14. ░░ Subject: Session 14 has been terminated ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A session with the ID 14 has been terminated. Sep 01 08:25:35 managed-node1 sshd[9953]: Accepted publickey for root from 10.31.9.95 port 56064 ssh2: RSA SHA256:9j1blwt3wcrRiGYZQ7ZGu9axm3cDklH6/z4c+Ee8CzE Sep 01 08:25:35 managed-node1 systemd-logind[606]: New session 15 of user root. ░░ Subject: A new session 15 has been created for user root ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ Documentation: sd-login(3) ░░ ░░ A new session with the ID 15 has been created for the user root. ░░ ░░ The leading process of the session is 9953. Sep 01 08:25:35 managed-node1 systemd[1]: Started Session 15 of User root. ░░ Subject: A start job for unit session-15.scope has finished successfully ░░ Defined-By: systemd ░░ Support: https://access.redhat.com/support ░░ ░░ A start job for unit session-15.scope has finished successfully. ░░ ░░ The job identifier is 1989. Sep 01 08:25:35 managed-node1 sshd[9953]: pam_unix(sshd:session): session opened for user root(uid=0) by root(uid=0)