[DEPRECATION WARNING]: ANSIBLE_COLLECTIONS_PATHS option, does not fit var naming standard, use the singular form ANSIBLE_COLLECTIONS_PATH instead. This feature will be removed from ansible-core in version 2.19. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. 9704 1726776646.85275: starting run ansible-playbook [core 2.16.11] config file = None configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.12/site-packages/ansible ansible collection location = /tmp/collections-uMf executable location = /usr/local/bin/ansible-playbook python version = 3.12.1 (main, Feb 21 2024, 14:18:26) [GCC 8.5.0 20210514 (Red Hat 8.5.0-21)] (/usr/bin/python3.12) jinja version = 3.1.4 libyaml = True No config file found; using defaults 9704 1726776646.85737: Added group all to inventory 9704 1726776646.85739: Added group ungrouped to inventory 9704 1726776646.85743: Group all now contains ungrouped 9704 1726776646.85746: Examining possible inventory source: /tmp/kernel_settings-iny/inventory.yml 9704 1726776646.98044: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/cache 9704 1726776646.98089: Loading CacheModule 'memory' from /usr/local/lib/python3.12/site-packages/ansible/plugins/cache/memory.py 9704 1726776646.98124: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory 9704 1726776646.98185: Loading InventoryModule 'host_list' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/host_list.py 9704 1726776646.98257: Loaded config def from plugin (inventory/script) 9704 1726776646.98259: Loading InventoryModule 'script' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/script.py 9704 1726776646.98297: Loading InventoryModule 'auto' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/auto.py 9704 1726776646.98375: Loaded config def from plugin (inventory/yaml) 9704 1726776646.98378: Loading InventoryModule 'yaml' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/yaml.py 9704 1726776646.98467: Loading InventoryModule 'ini' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/ini.py 9704 1726776646.98894: Loading InventoryModule 'toml' from /usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/toml.py 9704 1726776646.98896: Attempting to use plugin host_list (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/host_list.py) 9704 1726776646.98898: Attempting to use plugin script (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/script.py) 9704 1726776646.98902: Attempting to use plugin auto (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/auto.py) 9704 1726776646.98906: Loading data from /tmp/kernel_settings-iny/inventory.yml 9704 1726776646.98952: /tmp/kernel_settings-iny/inventory.yml was not parsable by auto 9704 1726776646.98994: Attempting to use plugin yaml (/usr/local/lib/python3.12/site-packages/ansible/plugins/inventory/yaml.py) 9704 1726776646.99024: Loading data from /tmp/kernel_settings-iny/inventory.yml 9704 1726776646.99083: group all already in inventory 9704 1726776646.99088: set inventory_file for managed_node1 9704 1726776646.99091: set inventory_dir for managed_node1 9704 1726776646.99091: Added host managed_node1 to inventory 9704 1726776646.99093: Added host managed_node1 to group all 9704 1726776646.99093: set ansible_host for managed_node1 9704 1726776646.99094: set ansible_ssh_extra_args for managed_node1 9704 1726776646.99095: set inventory_file for managed_node2 9704 1726776646.99097: set inventory_dir for managed_node2 9704 1726776646.99097: Added host managed_node2 to inventory 9704 1726776646.99098: Added host managed_node2 to group all 9704 1726776646.99099: set ansible_host for managed_node2 9704 1726776646.99099: set ansible_ssh_extra_args for managed_node2 9704 1726776646.99100: set inventory_file for managed_node3 9704 1726776646.99102: set inventory_dir for managed_node3 9704 1726776646.99102: Added host managed_node3 to inventory 9704 1726776646.99103: Added host managed_node3 to group all 9704 1726776646.99104: set ansible_host for managed_node3 9704 1726776646.99104: set ansible_ssh_extra_args for managed_node3 9704 1726776646.99106: Reconcile groups and hosts in inventory. 9704 1726776646.99108: Group ungrouped now contains managed_node1 9704 1726776646.99109: Group ungrouped now contains managed_node2 9704 1726776646.99110: Group ungrouped now contains managed_node3 9704 1726776646.99166: '/usr/local/lib/python3.12/site-packages/ansible/plugins/vars/__init__' skipped due to reserved name 9704 1726776646.99249: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments 9704 1726776646.99281: Loading ModuleDocFragment 'vars_plugin_staging' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/vars_plugin_staging.py 9704 1726776646.99298: Loaded config def from plugin (vars/host_group_vars) 9704 1726776646.99300: Loading VarsModule 'host_group_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/vars/host_group_vars.py (found_in_cache=False, class_only=True) 9704 1726776646.99304: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/vars 9704 1726776646.99310: Loading VarsModule 'host_group_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/vars/host_group_vars.py (found_in_cache=True, class_only=False) 9704 1726776646.99338: Loading CacheModule 'memory' from /usr/local/lib/python3.12/site-packages/ansible/plugins/cache/memory.py (found_in_cache=True, class_only=False) 9704 1726776646.99571: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776646.99638: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py 9704 1726776646.99661: Loaded config def from plugin (connection/local) 9704 1726776646.99663: Loading Connection 'local' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/local.py (found_in_cache=False, class_only=True) 9704 1726776646.99994: Loaded config def from plugin (connection/paramiko_ssh) 9704 1726776646.99997: Loading Connection 'paramiko_ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/paramiko_ssh.py (found_in_cache=False, class_only=True) 9704 1726776647.00595: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 9704 1726776647.00619: Loaded config def from plugin (connection/psrp) 9704 1726776647.00620: Loading Connection 'psrp' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/psrp.py (found_in_cache=False, class_only=True) 9704 1726776647.01058: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 9704 1726776647.01081: Loaded config def from plugin (connection/ssh) 9704 1726776647.01083: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=False, class_only=True) 9704 1726776647.02261: Loading ModuleDocFragment 'connection_pipelining' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/connection_pipelining.py (found_in_cache=True, class_only=False) 9704 1726776647.02287: Loaded config def from plugin (connection/winrm) 9704 1726776647.02289: Loading Connection 'winrm' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/winrm.py (found_in_cache=False, class_only=True) 9704 1726776647.02313: '/usr/local/lib/python3.12/site-packages/ansible/plugins/shell/__init__' skipped due to reserved name 9704 1726776647.02358: Loading ModuleDocFragment 'shell_windows' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_windows.py 9704 1726776647.02398: Loaded config def from plugin (shell/cmd) 9704 1726776647.02399: Loading ShellModule 'cmd' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/cmd.py (found_in_cache=False, class_only=True) 9704 1726776647.02418: Loading ModuleDocFragment 'shell_windows' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_windows.py (found_in_cache=True, class_only=False) 9704 1726776647.02456: Loaded config def from plugin (shell/powershell) 9704 1726776647.02457: Loading ShellModule 'powershell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/powershell.py (found_in_cache=False, class_only=True) 9704 1726776647.02493: Loading ModuleDocFragment 'shell_common' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/shell_common.py 9704 1726776647.02597: Loaded config def from plugin (shell/sh) 9704 1726776647.02598: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=False, class_only=True) 9704 1726776647.02624: '/usr/local/lib/python3.12/site-packages/ansible/plugins/become/__init__' skipped due to reserved name 9704 1726776647.02697: Loaded config def from plugin (become/runas) 9704 1726776647.02698: Loading BecomeModule 'runas' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/runas.py (found_in_cache=False, class_only=True) 9704 1726776647.02808: Loaded config def from plugin (become/su) 9704 1726776647.02810: Loading BecomeModule 'su' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/su.py (found_in_cache=False, class_only=True) 9704 1726776647.02907: Loaded config def from plugin (become/sudo) 9704 1726776647.02908: Loading BecomeModule 'sudo' from /usr/local/lib/python3.12/site-packages/ansible/plugins/become/sudo.py (found_in_cache=False, class_only=True) running playbook inside collection fedora.linux_system_roles 9704 1726776647.02934: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml 9704 1726776647.03303: trying /usr/local/lib/python3.12/site-packages/ansible/modules 9704 1726776647.05270: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/action 9704 1726776647.05451: in VariableManager get_vars() 9704 1726776647.05466: done with get_vars() 9704 1726776647.05491: in VariableManager get_vars() 9704 1726776647.05500: done with get_vars() 9704 1726776647.05635: in VariableManager get_vars() 9704 1726776647.05645: done with get_vars() 9704 1726776647.05737: in VariableManager get_vars() 9704 1726776647.05745: done with get_vars() 9704 1726776647.05777: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/callback 9704 1726776647.05786: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__ redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug redirecting (type: callback) ansible.builtin.debug to ansible.posix.debug 9704 1726776647.05942: Loading ModuleDocFragment 'default_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/default_callback.py 9704 1726776647.06039: Loaded config def from plugin (callback/ansible_collections.ansible.posix.plugins.callback.debug) 9704 1726776647.06041: Loading CallbackModule 'ansible_collections.ansible.posix.plugins.callback.debug' from /tmp/collections-uMf/ansible_collections/ansible/posix/plugins/callback/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) 9704 1726776647.06062: '/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__init__' skipped due to reserved name 9704 1726776647.06079: Loading ModuleDocFragment 'default_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/default_callback.py (found_in_cache=True, class_only=False) 9704 1726776647.06178: Loading ModuleDocFragment 'result_format_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/result_format_callback.py 9704 1726776647.06217: Loaded config def from plugin (callback/default) 9704 1726776647.06218: Loading CallbackModule 'default' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/default.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 9704 1726776647.06948: Loaded config def from plugin (callback/junit) 9704 1726776647.06950: Loading CallbackModule 'junit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/junit.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 9704 1726776647.06982: Loading ModuleDocFragment 'result_format_callback' from /usr/local/lib/python3.12/site-packages/ansible/plugins/doc_fragments/result_format_callback.py (found_in_cache=True, class_only=False) 9704 1726776647.07021: Loaded config def from plugin (callback/minimal) 9704 1726776647.07022: Loading CallbackModule 'minimal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/minimal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 9704 1726776647.07052: Loading CallbackModule 'oneline' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/oneline.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) 9704 1726776647.07091: Loaded config def from plugin (callback/tree) 9704 1726776647.07092: Loading CallbackModule 'tree' from /usr/local/lib/python3.12/site-packages/ansible/plugins/callback/tree.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) redirecting (type: callback) ansible.builtin.profile_tasks to ansible.posix.profile_tasks 9704 1726776647.07173: Loaded config def from plugin (callback/ansible_collections.ansible.posix.plugins.callback.profile_tasks) 9704 1726776647.07175: Loading CallbackModule 'ansible_collections.ansible.posix.plugins.callback.profile_tasks' from /tmp/collections-uMf/ansible_collections/ansible/posix/plugins/callback/profile_tasks.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/callback:/usr/local/lib/python3.12/site-packages/ansible/plugins/callback/__pycache__) (found_in_cache=False, class_only=True) Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_simple_settings.yml ******************************************** 1 plays in /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml 9704 1726776647.07191: in VariableManager get_vars() 9704 1726776647.07199: done with get_vars() 9704 1726776647.07202: in VariableManager get_vars() 9704 1726776647.07208: done with get_vars() 9704 1726776647.07211: variable 'omit' from source: magic vars 9704 1726776647.07237: in VariableManager get_vars() 9704 1726776647.07246: done with get_vars() 9704 1726776647.07259: variable 'omit' from source: magic vars PLAY [Test simple kernel settings] ********************************************* 9704 1726776647.07645: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy 9704 1726776647.07691: Loading StrategyModule 'linear' from /usr/local/lib/python3.12/site-packages/ansible/plugins/strategy/linear.py 9704 1726776647.07716: getting the remaining hosts for this loop 9704 1726776647.07721: done getting the remaining hosts for this loop 9704 1726776647.07722: getting the next task for host managed_node3 9704 1726776647.07725: done getting next task for host managed_node3 9704 1726776647.07726: ^ task is: TASK: Gathering Facts 9704 1726776647.07727: ^ state is: HOST STATE: block=0, task=0, rescue=0, always=0, handlers=0, run_state=0, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=True, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776647.07730: getting variables 9704 1726776647.07731: in VariableManager get_vars() 9704 1726776647.07737: Calling all_inventory to load vars for managed_node3 9704 1726776647.07751: Calling groups_inventory to load vars for managed_node3 9704 1726776647.07752: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776647.07762: Calling all_plugins_play to load vars for managed_node3 9704 1726776647.07768: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776647.07770: Calling groups_plugins_play to load vars for managed_node3 9704 1726776647.07791: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776647.07824: done with get_vars() 9704 1726776647.07830: done getting variables 9704 1726776647.07874: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (found_in_cache=False, class_only=True) TASK [Gathering Facts] ********************************************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:2 Thursday 19 September 2024 16:10:47 -0400 (0:00:00.007) 0:00:00.007 **** 9704 1726776647.07888: entering _queue_task() for managed_node3/gather_facts 9704 1726776647.07889: Creating lock for gather_facts 9704 1726776647.08098: worker is 1 (out of 1 available) 9704 1726776647.08107: exiting _queue_task() for managed_node3/gather_facts 9704 1726776647.08119: done queuing things up, now waiting for results queue to drain 9704 1726776647.08121: waiting for pending results... 9717 1726776647.08196: running TaskExecutor() for managed_node3/TASK: Gathering Facts 9717 1726776647.08291: in run() - task 120fa90a-8a95-fee7-0cea-000000000014 9717 1726776647.08309: variable 'ansible_search_path' from source: unknown 9717 1726776647.08338: calling self._execute() 9717 1726776647.08379: variable 'ansible_host' from source: host vars for 'managed_node3' 9717 1726776647.08387: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9717 1726776647.08395: variable 'omit' from source: magic vars 9717 1726776647.08462: variable 'omit' from source: magic vars 9717 1726776647.08484: variable 'omit' from source: magic vars 9717 1726776647.08510: variable 'omit' from source: magic vars 9717 1726776647.08542: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9717 1726776647.08568: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9717 1726776647.08586: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9717 1726776647.08600: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9717 1726776647.08613: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9717 1726776647.08636: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9717 1726776647.08641: variable 'ansible_host' from source: host vars for 'managed_node3' 9717 1726776647.08645: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9717 1726776647.08713: Set connection var ansible_module_compression to ZIP_DEFLATED 9717 1726776647.08722: Set connection var ansible_shell_type to sh 9717 1726776647.08731: Set connection var ansible_timeout to 10 9717 1726776647.08737: Set connection var ansible_pipelining to False 9717 1726776647.08744: Set connection var ansible_shell_executable to /bin/sh 9717 1726776647.08747: Set connection var ansible_connection to ssh 9717 1726776647.08762: variable 'ansible_shell_executable' from source: unknown 9717 1726776647.08766: variable 'ansible_connection' from source: unknown 9717 1726776647.08769: variable 'ansible_module_compression' from source: unknown 9717 1726776647.08773: variable 'ansible_shell_type' from source: unknown 9717 1726776647.08776: variable 'ansible_shell_executable' from source: unknown 9717 1726776647.08780: variable 'ansible_host' from source: host vars for 'managed_node3' 9717 1726776647.08784: variable 'ansible_pipelining' from source: unknown 9717 1726776647.08788: variable 'ansible_timeout' from source: unknown 9717 1726776647.08792: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9717 1726776647.08907: Loading ActionModule 'gather_facts' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/gather_facts.py (found_in_cache=True, class_only=False) 9717 1726776647.08918: variable 'omit' from source: magic vars 9717 1726776647.08924: starting attempt loop 9717 1726776647.08928: running the handler 9717 1726776647.08941: variable 'ansible_facts' from source: unknown 9717 1726776647.08956: _low_level_execute_command(): starting 9717 1726776647.08964: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9717 1726776647.11281: stdout chunk (state=2): >>>/root <<< 9717 1726776647.11396: stderr chunk (state=3): >>><<< 9717 1726776647.11402: stdout chunk (state=3): >>><<< 9717 1726776647.11420: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9717 1726776647.11435: _low_level_execute_command(): starting 9717 1726776647.11441: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662 `" && echo ansible-tmp-1726776647.114274-9717-163307621185662="` echo /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662 `" ) && sleep 0' 9717 1726776647.13822: stdout chunk (state=2): >>>ansible-tmp-1726776647.114274-9717-163307621185662=/root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662 <<< 9717 1726776647.13947: stderr chunk (state=3): >>><<< 9717 1726776647.13953: stdout chunk (state=3): >>><<< 9717 1726776647.13965: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776647.114274-9717-163307621185662=/root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662 , stderr= 9717 1726776647.13983: variable 'ansible_module_compression' from source: unknown 9717 1726776647.14023: ANSIBALLZ: Using generic lock for ansible.legacy.setup 9717 1726776647.14029: ANSIBALLZ: Acquiring lock 9717 1726776647.14033: ANSIBALLZ: Lock acquired: 140007686719920 9717 1726776647.14039: ANSIBALLZ: Creating module 9717 1726776647.40351: ANSIBALLZ: Writing module into payload 9717 1726776647.40524: ANSIBALLZ: Writing module 9717 1726776647.40551: ANSIBALLZ: Renaming module 9717 1726776647.40559: ANSIBALLZ: Done creating module 9717 1726776647.40593: variable 'ansible_facts' from source: unknown 9717 1726776647.40600: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9717 1726776647.40609: _low_level_execute_command(): starting 9717 1726776647.40616: _low_level_execute_command(): executing: /bin/sh -c 'echo PLATFORM; uname; echo FOUND; command -v '"'"'python3.12'"'"'; command -v '"'"'python3.11'"'"'; command -v '"'"'python3.10'"'"'; command -v '"'"'python3.9'"'"'; command -v '"'"'python3.8'"'"'; command -v '"'"'python3.7'"'"'; command -v '"'"'python3.6'"'"'; command -v '"'"'/usr/bin/python3'"'"'; command -v '"'"'/usr/libexec/platform-python'"'"'; command -v '"'"'python2.7'"'"'; command -v '"'"'/usr/bin/python'"'"'; command -v '"'"'python'"'"'; echo ENDFOUND && sleep 0' 9717 1726776647.43434: stdout chunk (state=2): >>>PLATFORM Linux FOUND /usr/bin/python3.12 /usr/bin/python3.6 /usr/bin/python3 /usr/libexec/platform-python ENDFOUND <<< 9717 1726776647.43444: stderr chunk (state=2): >>><<< 9717 1726776647.43455: stdout chunk (state=3): >>><<< 9717 1726776647.43469: _low_level_execute_command() done: rc=0, stdout=PLATFORM Linux FOUND /usr/bin/python3.12 /usr/bin/python3.6 /usr/bin/python3 /usr/libexec/platform-python ENDFOUND , stderr= 9717 1726776647.43476 [managed_node3]: found interpreters: ['/usr/bin/python3.12', '/usr/bin/python3.6', '/usr/bin/python3', '/usr/libexec/platform-python'] 9717 1726776647.43520: _low_level_execute_command(): starting 9717 1726776647.43527: _low_level_execute_command(): executing: /bin/sh -c '/usr/bin/python3.12 && sleep 0' 9717 1726776647.44164: Sending initial data 9717 1726776647.44172: Sent initial data (1234 bytes) 9717 1726776647.47788: stdout chunk (state=3): >>>{"platform_dist_result": [], "osrelease_content": "NAME=\"CentOS Stream\"\nVERSION=\"8\"\nID=\"centos\"\nID_LIKE=\"rhel fedora\"\nVERSION_ID=\"8\"\nPLATFORM_ID=\"platform:el8\"\nPRETTY_NAME=\"CentOS Stream 8\"\nANSI_COLOR=\"0;31\"\nCPE_NAME=\"cpe:/o:centos:centos:8\"\nHOME_URL=\"https://centos.org/\"\nBUG_REPORT_URL=\"https://bugzilla.redhat.com/\"\nREDHAT_SUPPORT_PRODUCT=\"Red Hat Enterprise Linux 8\"\nREDHAT_SUPPORT_PRODUCT_VERSION=\"CentOS Stream\"\n"} <<< 9717 1726776647.48260: stderr chunk (state=3): >>><<< 9717 1726776647.48268: stdout chunk (state=3): >>><<< 9717 1726776647.48283: _low_level_execute_command() done: rc=0, stdout={"platform_dist_result": [], "osrelease_content": "NAME=\"CentOS Stream\"\nVERSION=\"8\"\nID=\"centos\"\nID_LIKE=\"rhel fedora\"\nVERSION_ID=\"8\"\nPLATFORM_ID=\"platform:el8\"\nPRETTY_NAME=\"CentOS Stream 8\"\nANSI_COLOR=\"0;31\"\nCPE_NAME=\"cpe:/o:centos:centos:8\"\nHOME_URL=\"https://centos.org/\"\nBUG_REPORT_URL=\"https://bugzilla.redhat.com/\"\nREDHAT_SUPPORT_PRODUCT=\"Red Hat Enterprise Linux 8\"\nREDHAT_SUPPORT_PRODUCT_VERSION=\"CentOS Stream\"\n"} , stderr= 9717 1726776647.48327: variable 'ansible_facts' from source: unknown 9717 1726776647.48335: variable 'ansible_facts' from source: unknown 9717 1726776647.48346: variable 'ansible_module_compression' from source: unknown 9717 1726776647.48375: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.setup-ZIP_DEFLATED 9717 1726776647.48401: variable 'ansible_facts' from source: unknown 9717 1726776647.48546: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/AnsiballZ_setup.py 9717 1726776647.48649: Sending initial data 9717 1726776647.48657: Sent initial data (151 bytes) 9717 1726776647.51417: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpzvpzqj2l /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/AnsiballZ_setup.py <<< 9717 1726776647.53710: stderr chunk (state=3): >>><<< 9717 1726776647.53718: stdout chunk (state=3): >>><<< 9717 1726776647.53747: done transferring module to remote 9717 1726776647.53759: _low_level_execute_command(): starting 9717 1726776647.53766: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/ /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/AnsiballZ_setup.py && sleep 0' 9717 1726776647.56353: stderr chunk (state=2): >>><<< 9717 1726776647.56366: stdout chunk (state=2): >>><<< 9717 1726776647.56382: _low_level_execute_command() done: rc=0, stdout=, stderr= 9717 1726776647.56387: _low_level_execute_command(): starting 9717 1726776647.56393: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/AnsiballZ_setup.py && sleep 0' 9717 1726776648.17122: stdout chunk (state=2): >>> {"ansible_facts": {"ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAN9/yiwbHBqp6BhrLpcFqfxmaR+Gydfza4qvQePCqshEAYkkgSwyGPM48Gqa1wGCo9BjeQE1w2XHWjuuIIca+e/e733nN4lgIv9KkPTlQTV5gUs8mDS3KgjLjf+EFbnq+NrzrsuLeTr3FxAe1SkwhTtzR8Ru+q4cwxuQfMlNUVHDAAAAFQCUlvg4ja1xM8kP9hsKzR3Xp4D+NwAAAIEAzGBdxMNLb9i61ovOes9uvtohXJL9hWLt2QCcY0P6n8DFwqqP82nztvsJoAKvW3cdN8WdTAzivavEUJ7jVNBx/yrUk7e8S0d92cyM3OK88vkAQ6w1Y7E19iS1vtv8r/y5UN45V+5+YunQnHb+ClAql8b0SX0U04tXo12BaLs49rkAAACAKx4t74PGIm29HwZzvBWyPWgZPOgmXaesymHzj6CkyTXW6Es2Oo6M/N2Aw/51UsKkQT1O0M+ejY6SoLOzvcmkJvblGG2BnKiVh8kCjG+if13CTJxYY1I6wUtP6gzmj8Fjs+bvWetj2lyV/NpCjx1O8LuL2qQsz2ly0VbPWWOVe00=", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC5qW6qiwAUTS5zoCVLjcFBLwdbX9/fkWFMSlr9xWP1j6wqRwfzwDW9czu9zq44FZFdkGL6k48Q3dDzYwMljHztPoSvopeS0B7RQfyIJL5+9XmM5r1qAb8ju/VSuivIZrdF6D/vH1KuACJF/t0plxcRWXZHgmC+HSJpyu765eEdvWCHxbD8/Q3iPgwJv7uDsJ2qS+aI1WxCdx+2PSD6xc0vrK6Rskhqd12UGISCqFitkrWJpzSSOvt2Y1KbvA7oMH38vJD0KdiW/RBxdRTHPCImc9FE12mIC9XfyjpIAepPaKq9sKZwuq4hlfZj+Dr8cA+ReD4ZcMvtuU08NM6Z64VTYRU5wKZxYqafEhiZ6mjFnt3hv/SeGY36j4pw6liEL2Bn/iJXJV/Ifo/vNxhxuupibPVT8OMy/q5s9WUfFLZjZNVjjGo2P8SjPMk6FRZTjsdXHuuqI1Lub0C12+K3Huf6JqG0tKPbse8RBIdEeELY8QgPvBt9U7HFM6err71crJE=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAJmNXuRva4knSO/2PkZz2qXhfiSGml+fo63aWeY3RqdZZA6LNawbU67tSaAsJgvMHE7P6GFz/EGpS6tIHBHEDE=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIGgUUw4U1wlo0yxtM7k1xHct5ZV/1CLZTY/eTAaPNO5N", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_system": "Linux", "ansible_kernel": "4.18.0-553.5.1.el8.x86_64", "ansible_kernel_version": "#1 SMP Tue May 21 05:46:01 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.6.8", "ansible_fqdn": "ip-10-31-8-186.us-east-1.aws.redhat.com", "ansible_hostname": "ip-10-31-8-18<<< 9717 1726776648.17162: stdout chunk (state=3): >>>6", "ansible_nodename": "ip-10-31-8-186.us-east-1.aws.redhat.com", "ansible_domain": "us-east-1.aws.redhat.com", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "1897f356ead34fc79bf09198c525f66f", "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_python": {"version": {"major": 3, "minor": 6, "micro": 8, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0], "executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython"}, "ansible_is_chroot": false, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "8", "ansible_distribution_major_version": "8", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_hostnqn": "", "ansible_local": {}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_iscsi_iqn": "", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-553.5.1.el8.x86_64", "root": "UUID=fe591198-9082-4b15-9b62-e83518524cd2", "ro": true, "crashkernel": "auto", "net.ifnames": "0", "rhgb": true, "quiet": true}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-553.5.1.el8.x86_64", "root": "UUID=fe591198-9082-4b15-9b62-e83518524cd2", "ro": true, "crashkernel": "auto", "net.ifnames": "0", "rhgb": true, "quiet": true}, "ansible_fibre_channel_wwn": [], "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Thursday", "weekday_number": "4", "weeknumber": "38", "day": "19", "hour": "16", "minute": "10", "second": "47", "epoch": "1726776647", "epoch_int": "1726776647", "date": "2024-09-19", "time": "16:10:47", "iso8601_micro": "2024-09-19T20:10:47.868216Z", "iso8601": "2024-09-19T20:10:47Z", "iso8601_basic": "20240919T161047868216", "iso8601_basic_short": "20240919T161047", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_pkg_mgr": "dnf", "ansible_fips": false, "ansible_loadavg": {"1m": 0.79, "5m": 0.38, "15m": 0.16}, "ansible_apparmor": {"status": "disabled"}, "ansible_lsb": {}, "ansible_env": {"LS_COLORS": "", "SSH_CONNECTION": "10.31.10.210 33518 10.31.8.186 22", "_": "/usr/libexec/platform-python", "LANG": "en_US.UTF-8", "which_declare": "declare -f", "XDG_SESSION_ID": "5", "USER": "root", "SELINUX_ROLE_REQUESTED": "", "PWD": "/root", "HOME": "/root", "SSH_CLIENT": "10.31.10.210 33518 22", "SELINUX_LEVEL_REQUESTED": "", "SSH_TTY": "/dev/pts/0", "SHELL": "/bin/bash", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "2", "LOGNAME": "root", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "XDG_RUNTIME_DIR": "/run/user/0", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_npro<<< 9717 1726776648.17195: stdout chunk (state=3): >>>c": 2, "ansible_memtotal_mb": 3539, "ansible_memfree_mb": 2697, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3539, "used": 842, "free": 2697}, "nocache": {"free": 3295, "used": 244}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec27856d-4839-62f6-ec8d-d42383c2af48", "ansible_product_uuid": "ec27856d-4839-62f6-ec8d-d42383c2af48", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "0", "partitions": {"xvda1": {"links": {"ids": [], "uuids": ["fe591198-9082-4b15-9b62-e83518524cd2"], "labels": [], "masters": []}, "start": "2048", "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "uuid": "fe591198-9082-4b15-9b62-e83518524cd2", "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda1": ["fe591198-9082-4b15-9b62-e83518524cd2"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 255, "ansible_lvm": "N/A", "ansible_mounts": [{"mount": "/", "device": "/dev/xvda1", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_total": 268423901184, "size_available": 263514820608, "block_size": 4096, "block_total": 65533179, "block_available": 64334673, "block_used": 1198506, "inode_total": 131071472, "inode_available": 130994299, "inode_used": 77173, "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"}], "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "12:0d:b7:e6:b2:93", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.8.186", "broadcast": "10.31.11.255", "netmask": "255.255.252.0", "network": "10.31.8.0", "prefix": "22"}, "ipv6": [{"address": "fe80::100d:b7ff:fee6:b293", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_l<<< 9717 1726776648.17218: stdout chunk (state=3): >>>ist": "off [fixed]", "rx_udp_gro_forwarding": "off", "rx_gro_list": "off", "tls_hw_rx_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "rx_udp_gro_forwarding": "off", "rx_gro_list": "off", "tls_hw_rx_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.8.1", "interface": "eth0", "address": "10.31.8.186", "broadcast": "10.31.11.255", "netmask": "255.255.252.0", "network": "10.31.8.0", "prefix": "22", "macaddress": "12:0d:b7:e6:b2:93", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.8.186"], "ansible_all_ipv6_addresses": ["fe80::100d:b7ff:fee6:b293"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.8.186", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::100d:b7ff:fee6:b293"]}, "ansible_service_mgr": "systemd", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} <<< 9717 1726776648.18911: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9717 1726776648.18920: stdout chunk (state=3): >>><<< 9717 1726776648.18933: stderr chunk (state=3): >>><<< 9717 1726776648.18961: _low_level_execute_command() done: rc=0, stdout= {"ansible_facts": {"ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAN9/yiwbHBqp6BhrLpcFqfxmaR+Gydfza4qvQePCqshEAYkkgSwyGPM48Gqa1wGCo9BjeQE1w2XHWjuuIIca+e/e733nN4lgIv9KkPTlQTV5gUs8mDS3KgjLjf+EFbnq+NrzrsuLeTr3FxAe1SkwhTtzR8Ru+q4cwxuQfMlNUVHDAAAAFQCUlvg4ja1xM8kP9hsKzR3Xp4D+NwAAAIEAzGBdxMNLb9i61ovOes9uvtohXJL9hWLt2QCcY0P6n8DFwqqP82nztvsJoAKvW3cdN8WdTAzivavEUJ7jVNBx/yrUk7e8S0d92cyM3OK88vkAQ6w1Y7E19iS1vtv8r/y5UN45V+5+YunQnHb+ClAql8b0SX0U04tXo12BaLs49rkAAACAKx4t74PGIm29HwZzvBWyPWgZPOgmXaesymHzj6CkyTXW6Es2Oo6M/N2Aw/51UsKkQT1O0M+ejY6SoLOzvcmkJvblGG2BnKiVh8kCjG+if13CTJxYY1I6wUtP6gzmj8Fjs+bvWetj2lyV/NpCjx1O8LuL2qQsz2ly0VbPWWOVe00=", "ansible_ssh_host_key_dsa_public_keytype": "ssh-dss", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABgQC5qW6qiwAUTS5zoCVLjcFBLwdbX9/fkWFMSlr9xWP1j6wqRwfzwDW9czu9zq44FZFdkGL6k48Q3dDzYwMljHztPoSvopeS0B7RQfyIJL5+9XmM5r1qAb8ju/VSuivIZrdF6D/vH1KuACJF/t0plxcRWXZHgmC+HSJpyu765eEdvWCHxbD8/Q3iPgwJv7uDsJ2qS+aI1WxCdx+2PSD6xc0vrK6Rskhqd12UGISCqFitkrWJpzSSOvt2Y1KbvA7oMH38vJD0KdiW/RBxdRTHPCImc9FE12mIC9XfyjpIAepPaKq9sKZwuq4hlfZj+Dr8cA+ReD4ZcMvtuU08NM6Z64VTYRU5wKZxYqafEhiZ6mjFnt3hv/SeGY36j4pw6liEL2Bn/iJXJV/Ifo/vNxhxuupibPVT8OMy/q5s9WUfFLZjZNVjjGo2P8SjPMk6FRZTjsdXHuuqI1Lub0C12+K3Huf6JqG0tKPbse8RBIdEeELY8QgPvBt9U7HFM6err71crJE=", "ansible_ssh_host_key_rsa_public_keytype": "ssh-rsa", "ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAJmNXuRva4knSO/2PkZz2qXhfiSGml+fo63aWeY3RqdZZA6LNawbU67tSaAsJgvMHE7P6GFz/EGpS6tIHBHEDE=", "ansible_ssh_host_key_ecdsa_public_keytype": "ecdsa-sha2-nistp256", "ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIGgUUw4U1wlo0yxtM7k1xHct5ZV/1CLZTY/eTAaPNO5N", "ansible_ssh_host_key_ed25519_public_keytype": "ssh-ed25519", "ansible_system": "Linux", "ansible_kernel": "4.18.0-553.5.1.el8.x86_64", "ansible_kernel_version": "#1 SMP Tue May 21 05:46:01 UTC 2024", "ansible_machine": "x86_64", "ansible_python_version": "3.6.8", "ansible_fqdn": "ip-10-31-8-186.us-east-1.aws.redhat.com", "ansible_hostname": "ip-10-31-8-186", "ansible_nodename": "ip-10-31-8-186.us-east-1.aws.redhat.com", "ansible_domain": "us-east-1.aws.redhat.com", "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_userspace_architecture": "x86_64", "ansible_machine_id": "1897f356ead34fc79bf09198c525f66f", "ansible_virtualization_type": "xen", "ansible_virtualization_role": "guest", "ansible_virtualization_tech_guest": ["xen"], "ansible_virtualization_tech_host": [], "ansible_python": {"version": {"major": 3, "minor": 6, "micro": 8, "releaselevel": "final", "serial": 0}, "version_info": [3, 6, 8, "final", 0], "executable": "/usr/libexec/platform-python", "has_sslcontext": true, "type": "cpython"}, "ansible_is_chroot": false, "ansible_distribution": "CentOS", "ansible_distribution_release": "Stream", "ansible_distribution_version": "8", "ansible_distribution_major_version": "8", "ansible_distribution_file_path": "/etc/centos-release", "ansible_distribution_file_variety": "CentOS", "ansible_distribution_file_parsed": true, "ansible_os_family": "RedHat", "ansible_user_id": "root", "ansible_user_uid": 0, "ansible_user_gid": 0, "ansible_user_gecos": "root", "ansible_user_dir": "/root", "ansible_user_shell": "/bin/bash", "ansible_real_user_id": 0, "ansible_effective_user_id": 0, "ansible_real_group_id": 0, "ansible_effective_group_id": 0, "ansible_hostnqn": "", "ansible_local": {}, "ansible_selinux_python_present": true, "ansible_selinux": {"status": "enabled", "policyvers": 33, "config_mode": "enforcing", "mode": "enforcing", "type": "targeted"}, "ansible_iscsi_iqn": "", "ansible_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-553.5.1.el8.x86_64", "root": "UUID=fe591198-9082-4b15-9b62-e83518524cd2", "ro": true, "crashkernel": "auto", "net.ifnames": "0", "rhgb": true, "quiet": true}, "ansible_proc_cmdline": {"BOOT_IMAGE": "(hd0,msdos1)/boot/vmlinuz-4.18.0-553.5.1.el8.x86_64", "root": "UUID=fe591198-9082-4b15-9b62-e83518524cd2", "ro": true, "crashkernel": "auto", "net.ifnames": "0", "rhgb": true, "quiet": true}, "ansible_fibre_channel_wwn": [], "ansible_system_capabilities_enforced": "False", "ansible_system_capabilities": [], "ansible_dns": {"search": ["us-east-1.aws.redhat.com"], "nameservers": ["10.29.169.13", "10.29.170.12", "10.2.32.1"]}, "ansible_date_time": {"year": "2024", "month": "09", "weekday": "Thursday", "weekday_number": "4", "weeknumber": "38", "day": "19", "hour": "16", "minute": "10", "second": "47", "epoch": "1726776647", "epoch_int": "1726776647", "date": "2024-09-19", "time": "16:10:47", "iso8601_micro": "2024-09-19T20:10:47.868216Z", "iso8601": "2024-09-19T20:10:47Z", "iso8601_basic": "20240919T161047868216", "iso8601_basic_short": "20240919T161047", "tz": "EDT", "tz_dst": "EDT", "tz_offset": "-0400"}, "ansible_pkg_mgr": "dnf", "ansible_fips": false, "ansible_loadavg": {"1m": 0.79, "5m": 0.38, "15m": 0.16}, "ansible_apparmor": {"status": "disabled"}, "ansible_lsb": {}, "ansible_env": {"LS_COLORS": "", "SSH_CONNECTION": "10.31.10.210 33518 10.31.8.186 22", "_": "/usr/libexec/platform-python", "LANG": "en_US.UTF-8", "which_declare": "declare -f", "XDG_SESSION_ID": "5", "USER": "root", "SELINUX_ROLE_REQUESTED": "", "PWD": "/root", "HOME": "/root", "SSH_CLIENT": "10.31.10.210 33518 22", "SELINUX_LEVEL_REQUESTED": "", "SSH_TTY": "/dev/pts/0", "SHELL": "/bin/bash", "SELINUX_USE_CURRENT_RANGE": "", "SHLVL": "2", "LOGNAME": "root", "DBUS_SESSION_BUS_ADDRESS": "unix:path=/run/user/0/bus", "XDG_RUNTIME_DIR": "/run/user/0", "PATH": "/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin", "DEBUGINFOD_URLS": "https://debuginfod.centos.org/ ", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "BASH_FUNC_which%%": "() { ( alias;\n eval ${which_declare} ) | /usr/bin/which --tty-only --read-alias --read-functions --show-tilde --show-dot $@\n}"}, "ansible_processor": ["0", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz", "1", "GenuineIntel", "Intel(R) Xeon(R) CPU E5-2666 v3 @ 2.90GHz"], "ansible_processor_count": 1, "ansible_processor_cores": 1, "ansible_processor_threads_per_core": 2, "ansible_processor_vcpus": 2, "ansible_processor_nproc": 2, "ansible_memtotal_mb": 3539, "ansible_memfree_mb": 2697, "ansible_swaptotal_mb": 0, "ansible_swapfree_mb": 0, "ansible_memory_mb": {"real": {"total": 3539, "used": 842, "free": 2697}, "nocache": {"free": 3295, "used": 244}, "swap": {"total": 0, "free": 0, "used": 0, "cached": 0}}, "ansible_bios_date": "08/24/2006", "ansible_bios_vendor": "Xen", "ansible_bios_version": "4.11.amazon", "ansible_board_asset_tag": "NA", "ansible_board_name": "NA", "ansible_board_serial": "NA", "ansible_board_vendor": "NA", "ansible_board_version": "NA", "ansible_chassis_asset_tag": "NA", "ansible_chassis_serial": "NA", "ansible_chassis_vendor": "Xen", "ansible_chassis_version": "NA", "ansible_form_factor": "Other", "ansible_product_name": "HVM domU", "ansible_product_serial": "ec27856d-4839-62f6-ec8d-d42383c2af48", "ansible_product_uuid": "ec27856d-4839-62f6-ec8d-d42383c2af48", "ansible_product_version": "4.11.amazon", "ansible_system_vendor": "Xen", "ansible_devices": {"xvda": {"virtual": 1, "links": {"ids": [], "uuids": [], "labels": [], "masters": []}, "vendor": null, "model": null, "sas_address": null, "sas_device_handle": null, "removable": "0", "support_discard": "0", "partitions": {"xvda1": {"links": {"ids": [], "uuids": ["fe591198-9082-4b15-9b62-e83518524cd2"], "labels": [], "masters": []}, "start": "2048", "sectors": "524285919", "sectorsize": 512, "size": "250.00 GB", "uuid": "fe591198-9082-4b15-9b62-e83518524cd2", "holders": []}}, "rotational": "0", "scheduler_mode": "mq-deadline", "sectors": "524288000", "sectorsize": "512", "size": "250.00 GB", "host": "", "holders": []}}, "ansible_device_links": {"ids": {}, "uuids": {"xvda1": ["fe591198-9082-4b15-9b62-e83518524cd2"]}, "labels": {}, "masters": {}}, "ansible_uptime_seconds": 255, "ansible_lvm": "N/A", "ansible_mounts": [{"mount": "/", "device": "/dev/xvda1", "fstype": "xfs", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_total": 268423901184, "size_available": 263514820608, "block_size": 4096, "block_total": 65533179, "block_available": 64334673, "block_used": 1198506, "inode_total": 131071472, "inode_available": 130994299, "inode_used": 77173, "uuid": "fe591198-9082-4b15-9b62-e83518524cd2"}], "ansible_interfaces": ["lo", "eth0"], "ansible_eth0": {"device": "eth0", "macaddress": "12:0d:b7:e6:b2:93", "mtu": 9001, "active": true, "module": "xen_netfront", "type": "ether", "pciid": "vif-0", "promisc": false, "ipv4": {"address": "10.31.8.186", "broadcast": "10.31.11.255", "netmask": "255.255.252.0", "network": "10.31.8.0", "prefix": "22"}, "ipv6": [{"address": "fe80::100d:b7ff:fee6:b293", "prefix": "64", "scope": "link"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "on [fixed]", "tx_checksum_ip_generic": "off [fixed]", "tx_checksum_ipv6": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on", "tx_scatter_gather_fraglist": "off [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "off [fixed]", "tx_tcp_mangleid_segmentation": "off", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "off [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "off [fixed]", "tx_lockless": "off [fixed]", "netns_local": "off [fixed]", "tx_gso_robust": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "off [fixed]", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "off [fixed]", "tx_gso_list": "off [fixed]", "rx_udp_gro_forwarding": "off", "rx_gro_list": "off", "tls_hw_rx_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off", "loopback": "off [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_lo": {"device": "lo", "mtu": 65536, "active": true, "type": "loopback", "promisc": false, "ipv4": {"address": "127.0.0.1", "broadcast": "", "netmask": "255.0.0.0", "network": "127.0.0.0", "prefix": "8"}, "ipv6": [{"address": "::1", "prefix": "128", "scope": "host"}], "features": {"rx_checksumming": "on [fixed]", "tx_checksumming": "on", "tx_checksum_ipv4": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_checksum_ipv6": "off [fixed]", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_sctp": "on [fixed]", "scatter_gather": "on", "tx_scatter_gather": "on [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tcp_segmentation_offload": "on", "tx_tcp_segmentation": "on", "tx_tcp_ecn_segmentation": "on", "tx_tcp_mangleid_segmentation": "on", "tx_tcp6_segmentation": "on", "generic_segmentation_offload": "on", "generic_receive_offload": "on", "large_receive_offload": "off [fixed]", "rx_vlan_offload": "off [fixed]", "tx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "receive_hashing": "off [fixed]", "highdma": "on [fixed]", "rx_vlan_filter": "off [fixed]", "vlan_challenged": "on [fixed]", "tx_lockless": "on [fixed]", "netns_local": "on [fixed]", "tx_gso_robust": "off [fixed]", "tx_fcoe_segmentation": "off [fixed]", "tx_gre_segmentation": "off [fixed]", "tx_gre_csum_segmentation": "off [fixed]", "tx_ipxip4_segmentation": "off [fixed]", "tx_ipxip6_segmentation": "off [fixed]", "tx_udp_tnl_segmentation": "off [fixed]", "tx_udp_tnl_csum_segmentation": "off [fixed]", "tx_gso_partial": "off [fixed]", "tx_tunnel_remcsum_segmentation": "off [fixed]", "tx_sctp_segmentation": "on", "tx_esp_segmentation": "off [fixed]", "tx_udp_segmentation": "on", "tx_gso_list": "on", "rx_udp_gro_forwarding": "off", "rx_gro_list": "off", "tls_hw_rx_offload": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_nocache_copy": "off [fixed]", "loopback": "on [fixed]", "rx_fcs": "off [fixed]", "rx_all": "off [fixed]", "tx_vlan_stag_hw_insert": "off [fixed]", "rx_vlan_stag_hw_parse": "off [fixed]", "rx_vlan_stag_filter": "off [fixed]", "l2_fwd_offload": "off [fixed]", "hw_tc_offload": "off [fixed]", "esp_hw_offload": "off [fixed]", "esp_tx_csum_hw_offload": "off [fixed]", "rx_udp_tunnel_port_offload": "off [fixed]", "tls_hw_tx_offload": "off [fixed]", "rx_gro_hw": "off [fixed]", "tls_hw_record": "off [fixed]"}, "timestamping": [], "hw_timestamp_filters": []}, "ansible_default_ipv4": {"gateway": "10.31.8.1", "interface": "eth0", "address": "10.31.8.186", "broadcast": "10.31.11.255", "netmask": "255.255.252.0", "network": "10.31.8.0", "prefix": "22", "macaddress": "12:0d:b7:e6:b2:93", "mtu": 9001, "type": "ether", "alias": "eth0"}, "ansible_default_ipv6": {}, "ansible_all_ipv4_addresses": ["10.31.8.186"], "ansible_all_ipv6_addresses": ["fe80::100d:b7ff:fee6:b293"], "ansible_locally_reachable_ips": {"ipv4": ["10.31.8.186", "127.0.0.0/8", "127.0.0.1"], "ipv6": ["::1", "fe80::100d:b7ff:fee6:b293"]}, "ansible_service_mgr": "systemd", "gather_subset": ["all"], "module_setup": true}, "invocation": {"module_args": {"gather_subset": ["all"], "gather_timeout": 10, "filter": [], "fact_path": "/etc/ansible/facts.d"}}} , stderr=Shared connection to 10.31.8.186 closed. 9717 1726776648.19964: done with _execute_module (ansible.legacy.setup, {'_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.setup', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9717 1726776648.19989: _low_level_execute_command(): starting 9717 1726776648.19995: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776647.114274-9717-163307621185662/ > /dev/null 2>&1 && sleep 0' 9717 1726776648.22414: stderr chunk (state=2): >>><<< 9717 1726776648.22422: stdout chunk (state=2): >>><<< 9717 1726776648.22437: _low_level_execute_command() done: rc=0, stdout=, stderr= 9717 1726776648.22446: handler run complete 9717 1726776648.22515: variable 'ansible_facts' from source: unknown 9717 1726776648.22576: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9717 1726776648.22750: variable 'ansible_facts' from source: unknown 9717 1726776648.22811: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9717 1726776648.22883: attempt loop complete, returning result 9717 1726776648.22888: _execute() done 9717 1726776648.22891: dumping result to json 9717 1726776648.22906: done dumping result, returning 9717 1726776648.22911: done running TaskExecutor() for managed_node3/TASK: Gathering Facts [120fa90a-8a95-fee7-0cea-000000000014] 9717 1726776648.22915: sending task result for task 120fa90a-8a95-fee7-0cea-000000000014 9717 1726776648.23031: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000014 9717 1726776648.23036: WORKER PROCESS EXITING ok: [managed_node3] 9704 1726776648.25017: no more pending results, returning what we have 9704 1726776648.25019: results queue empty 9704 1726776648.25020: checking for any_errors_fatal 9704 1726776648.25021: done checking for any_errors_fatal 9704 1726776648.25021: checking for max_fail_percentage 9704 1726776648.25022: done checking for max_fail_percentage 9704 1726776648.25022: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.25023: done checking to see if all hosts have failed 9704 1726776648.25024: getting the remaining hosts for this loop 9704 1726776648.25025: done getting the remaining hosts for this loop 9704 1726776648.25027: getting the next task for host managed_node3 9704 1726776648.25031: done getting next task for host managed_node3 9704 1726776648.25032: ^ task is: TASK: meta (flush_handlers) 9704 1726776648.25034: ^ state is: HOST STATE: block=1, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.25036: getting variables 9704 1726776648.25036: in VariableManager get_vars() 9704 1726776648.25046: Calling all_inventory to load vars for managed_node3 9704 1726776648.25048: Calling groups_inventory to load vars for managed_node3 9704 1726776648.25049: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.25053: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.25054: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.25056: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.25156: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.25262: done with get_vars() 9704 1726776648.25269: done getting variables 9704 1726776648.25299: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__ 9704 1726776648.25338: in VariableManager get_vars() 9704 1726776648.25344: Calling all_inventory to load vars for managed_node3 9704 1726776648.25345: Calling groups_inventory to load vars for managed_node3 9704 1726776648.25347: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.25350: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.25352: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.25353: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.25435: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.25548: done with get_vars() 9704 1726776648.25557: done queuing things up, now waiting for results queue to drain 9704 1726776648.25558: results queue empty 9704 1726776648.25558: checking for any_errors_fatal 9704 1726776648.25561: done checking for any_errors_fatal 9704 1726776648.25561: checking for max_fail_percentage 9704 1726776648.25562: done checking for max_fail_percentage 9704 1726776648.25562: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.25562: done checking to see if all hosts have failed 9704 1726776648.25563: getting the remaining hosts for this loop 9704 1726776648.25563: done getting the remaining hosts for this loop 9704 1726776648.25565: getting the next task for host managed_node3 9704 1726776648.25568: done getting next task for host managed_node3 9704 1726776648.25569: ^ task is: TASK: Set platform independent vars used by this test 9704 1726776648.25570: ^ state is: HOST STATE: block=2, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.25571: getting variables 9704 1726776648.25572: in VariableManager get_vars() 9704 1726776648.25577: Calling all_inventory to load vars for managed_node3 9704 1726776648.25578: Calling groups_inventory to load vars for managed_node3 9704 1726776648.25579: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.25582: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.25583: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.25584: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.25661: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.25760: done with get_vars() 9704 1726776648.25766: done getting variables 9704 1726776648.25822: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Set platform independent vars used by this test] ************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:10 Thursday 19 September 2024 16:10:48 -0400 (0:00:01.179) 0:00:01.187 **** 9704 1726776648.25844: entering _queue_task() for managed_node3/include_vars 9704 1726776648.25845: Creating lock for include_vars 9704 1726776648.26015: worker is 1 (out of 1 available) 9704 1726776648.26030: exiting _queue_task() for managed_node3/include_vars 9704 1726776648.26040: done queuing things up, now waiting for results queue to drain 9704 1726776648.26042: waiting for pending results... 9759 1726776648.26127: running TaskExecutor() for managed_node3/TASK: Set platform independent vars used by this test 9759 1726776648.26214: in run() - task 120fa90a-8a95-fee7-0cea-000000000006 9759 1726776648.26226: variable 'ansible_search_path' from source: unknown 9759 1726776648.26254: calling self._execute() 9759 1726776648.26300: variable 'ansible_host' from source: host vars for 'managed_node3' 9759 1726776648.26309: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9759 1726776648.26316: variable 'omit' from source: magic vars 9759 1726776648.26386: variable 'omit' from source: magic vars 9759 1726776648.26409: variable 'omit' from source: magic vars 9759 1726776648.26433: variable 'omit' from source: magic vars 9759 1726776648.26462: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9759 1726776648.26491: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9759 1726776648.26512: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9759 1726776648.26524: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9759 1726776648.26537: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9759 1726776648.26567: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9759 1726776648.26573: variable 'ansible_host' from source: host vars for 'managed_node3' 9759 1726776648.26577: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9759 1726776648.26649: Set connection var ansible_module_compression to ZIP_DEFLATED 9759 1726776648.26658: Set connection var ansible_shell_type to sh 9759 1726776648.26666: Set connection var ansible_timeout to 10 9759 1726776648.26671: Set connection var ansible_pipelining to False 9759 1726776648.26678: Set connection var ansible_shell_executable to /bin/sh 9759 1726776648.26681: Set connection var ansible_connection to ssh 9759 1726776648.26697: variable 'ansible_shell_executable' from source: unknown 9759 1726776648.26703: variable 'ansible_connection' from source: unknown 9759 1726776648.26706: variable 'ansible_module_compression' from source: unknown 9759 1726776648.26712: variable 'ansible_shell_type' from source: unknown 9759 1726776648.26715: variable 'ansible_shell_executable' from source: unknown 9759 1726776648.26719: variable 'ansible_host' from source: host vars for 'managed_node3' 9759 1726776648.26723: variable 'ansible_pipelining' from source: unknown 9759 1726776648.26727: variable 'ansible_timeout' from source: unknown 9759 1726776648.26732: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9759 1726776648.26824: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9759 1726776648.26838: variable 'omit' from source: magic vars 9759 1726776648.26843: starting attempt loop 9759 1726776648.26847: running the handler 9759 1726776648.26857: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings 9759 1726776648.26875: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/vars/vars_simple_settings.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/vars_simple_settings.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/vars/vars_simple_settings.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/vars_simple_settings.yml 9759 1726776648.27039: handler run complete 9759 1726776648.27050: attempt loop complete, returning result 9759 1726776648.27054: _execute() done 9759 1726776648.27057: dumping result to json 9759 1726776648.27063: done dumping result, returning 9759 1726776648.27069: done running TaskExecutor() for managed_node3/TASK: Set platform independent vars used by this test [120fa90a-8a95-fee7-0cea-000000000006] 9759 1726776648.27075: sending task result for task 120fa90a-8a95-fee7-0cea-000000000006 9759 1726776648.27102: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000006 9759 1726776648.27105: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_profile_file": "# File managed by Ansible - DO NOT EDIT\n[main]\nsummary = kernel settings\n[sysctl]\nfs.epoll.max_user_watches = 785592\nfs.file-max = 379724\n[vm]\ntransparent_hugepages = madvise\n[sysfs]\n/sys/kernel/debug/x86/ibrs_enabled = 0\n/sys/kernel/debug/x86/pti_enabled = 0\n/sys/kernel/debug/x86/retp_enabled = 0\n{% if __kernel_settings_blcmdline_value | d() %}\n[bootloader]\ncmdline = {{ __kernel_settings_blcmdline_value }}\n{% endif %}\n", "kernel_settings_sysctl": [ { "name": "fs.epoll.max_user_watches", "value": 785592 }, { "name": "fs.file-max", "value": 379724 }, { "name": "no.such.param", "state": "absent" } ], "kernel_settings_sysfs": [ { "name": "/sys/kernel/debug/x86/pti_enabled", "value": 0 }, { "name": "/sys/kernel/debug/x86/retp_enabled", "value": 0 }, { "name": "/sys/kernel/debug/x86/ibrs_enabled", "value": 0 }, { "name": "/sys/not/found", "state": "absent" } ], "kernel_settings_systemd_cpu_affinity": { "state": "absent" }, "kernel_settings_transparent_hugepages": "madvise", "kernel_settings_transparent_hugepages_defrag": { "state": "absent" } }, "ansible_included_var_files": [ "/tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/vars_simple_settings.yml" ], "changed": false } 9704 1726776648.27336: no more pending results, returning what we have 9704 1726776648.27338: results queue empty 9704 1726776648.27339: checking for any_errors_fatal 9704 1726776648.27340: done checking for any_errors_fatal 9704 1726776648.27340: checking for max_fail_percentage 9704 1726776648.27341: done checking for max_fail_percentage 9704 1726776648.27341: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.27342: done checking to see if all hosts have failed 9704 1726776648.27342: getting the remaining hosts for this loop 9704 1726776648.27343: done getting the remaining hosts for this loop 9704 1726776648.27345: getting the next task for host managed_node3 9704 1726776648.27348: done getting next task for host managed_node3 9704 1726776648.27349: ^ task is: TASK: Disable bootloader cmdline testing on Fedora 9704 1726776648.27350: ^ state is: HOST STATE: block=2, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.27352: getting variables 9704 1726776648.27353: in VariableManager get_vars() 9704 1726776648.27372: Calling all_inventory to load vars for managed_node3 9704 1726776648.27374: Calling groups_inventory to load vars for managed_node3 9704 1726776648.27376: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.27382: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.27383: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.27385: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.27476: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.27578: done with get_vars() 9704 1726776648.27585: done getting variables 9704 1726776648.27651: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Disable bootloader cmdline testing on Fedora] **************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:14 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.018) 0:00:01.205 **** 9704 1726776648.27669: entering _queue_task() for managed_node3/set_fact 9704 1726776648.27670: Creating lock for set_fact 9704 1726776648.27841: worker is 1 (out of 1 available) 9704 1726776648.27855: exiting _queue_task() for managed_node3/set_fact 9704 1726776648.27865: done queuing things up, now waiting for results queue to drain 9704 1726776648.27867: waiting for pending results... 9760 1726776648.27959: running TaskExecutor() for managed_node3/TASK: Disable bootloader cmdline testing on Fedora 9760 1726776648.28051: in run() - task 120fa90a-8a95-fee7-0cea-000000000007 9760 1726776648.28067: variable 'ansible_search_path' from source: unknown 9760 1726776648.28098: calling self._execute() 9760 1726776648.28151: variable 'ansible_host' from source: host vars for 'managed_node3' 9760 1726776648.28159: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9760 1726776648.28168: variable 'omit' from source: magic vars 9760 1726776648.28488: variable 'ansible_distribution' from source: facts 9760 1726776648.28510: Evaluated conditional (ansible_distribution == "Fedora"): False 9760 1726776648.28515: when evaluation is False, skipping this task 9760 1726776648.28518: _execute() done 9760 1726776648.28521: dumping result to json 9760 1726776648.28525: done dumping result, returning 9760 1726776648.28532: done running TaskExecutor() for managed_node3/TASK: Disable bootloader cmdline testing on Fedora [120fa90a-8a95-fee7-0cea-000000000007] 9760 1726776648.28540: sending task result for task 120fa90a-8a95-fee7-0cea-000000000007 9760 1726776648.28561: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000007 9760 1726776648.28566: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "ansible_distribution == \"Fedora\"", "skip_reason": "Conditional result was False" } 9704 1726776648.28671: no more pending results, returning what we have 9704 1726776648.28673: results queue empty 9704 1726776648.28674: checking for any_errors_fatal 9704 1726776648.28678: done checking for any_errors_fatal 9704 1726776648.28679: checking for max_fail_percentage 9704 1726776648.28680: done checking for max_fail_percentage 9704 1726776648.28681: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.28681: done checking to see if all hosts have failed 9704 1726776648.28682: getting the remaining hosts for this loop 9704 1726776648.28683: done getting the remaining hosts for this loop 9704 1726776648.28686: getting the next task for host managed_node3 9704 1726776648.28691: done getting next task for host managed_node3 9704 1726776648.28693: ^ task is: TASK: Apply the settings - call the role 9704 1726776648.28694: ^ state is: HOST STATE: block=2, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.28697: getting variables 9704 1726776648.28698: in VariableManager get_vars() 9704 1726776648.28721: Calling all_inventory to load vars for managed_node3 9704 1726776648.28724: Calling groups_inventory to load vars for managed_node3 9704 1726776648.28726: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.28736: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.28738: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.28742: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.28847: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.28974: done with get_vars() 9704 1726776648.28981: done getting variables TASK [Apply the settings - call the role] ************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:25 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.013) 0:00:01.219 **** 9704 1726776648.29043: entering _queue_task() for managed_node3/include_role 9704 1726776648.29045: Creating lock for include_role 9704 1726776648.29203: worker is 1 (out of 1 available) 9704 1726776648.29218: exiting _queue_task() for managed_node3/include_role 9704 1726776648.29231: done queuing things up, now waiting for results queue to drain 9704 1726776648.29233: waiting for pending results... 9761 1726776648.29318: running TaskExecutor() for managed_node3/TASK: Apply the settings - call the role 9761 1726776648.29405: in run() - task 120fa90a-8a95-fee7-0cea-000000000008 9761 1726776648.29422: variable 'ansible_search_path' from source: unknown 9761 1726776648.29451: calling self._execute() 9761 1726776648.29495: variable 'ansible_host' from source: host vars for 'managed_node3' 9761 1726776648.29500: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9761 1726776648.29506: variable 'omit' from source: magic vars 9761 1726776648.29576: _execute() done 9761 1726776648.29580: dumping result to json 9761 1726776648.29583: done dumping result, returning 9761 1726776648.29586: done running TaskExecutor() for managed_node3/TASK: Apply the settings - call the role [120fa90a-8a95-fee7-0cea-000000000008] 9761 1726776648.29592: sending task result for task 120fa90a-8a95-fee7-0cea-000000000008 9761 1726776648.29613: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000008 9761 1726776648.29615: WORKER PROCESS EXITING 9704 1726776648.29816: no more pending results, returning what we have 9704 1726776648.29819: in VariableManager get_vars() 9704 1726776648.29842: Calling all_inventory to load vars for managed_node3 9704 1726776648.29844: Calling groups_inventory to load vars for managed_node3 9704 1726776648.29846: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.29852: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.29853: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.29855: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.29953: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.30056: done with get_vars() 9704 1726776648.30061: variable 'ansible_search_path' from source: unknown 9704 1726776648.30214: variable 'omit' from source: magic vars 9704 1726776648.30232: variable 'omit' from source: magic vars 9704 1726776648.30242: variable 'omit' from source: magic vars 9704 1726776648.30244: we have included files to process 9704 1726776648.30245: generating all_blocks data 9704 1726776648.30245: done generating all_blocks data 9704 1726776648.30246: processing included file: fedora.linux_system_roles.kernel_settings 9704 1726776648.30260: in VariableManager get_vars() 9704 1726776648.30268: done with get_vars() 9704 1726776648.30314: in VariableManager get_vars() 9704 1726776648.30324: done with get_vars() 9704 1726776648.30352: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/vars/main.yml 9704 1726776648.30456: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/defaults/main.yml 9704 1726776648.30492: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/meta/main.yml 9704 1726776648.30571: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml 9704 1726776648.31043: in VariableManager get_vars() 9704 1726776648.31055: done with get_vars() 9704 1726776648.31965: in VariableManager get_vars() 9704 1726776648.31977: done with get_vars() 9704 1726776648.32085: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/handlers/main.yml 9704 1726776648.32502: iterating over new_blocks loaded from include file 9704 1726776648.32504: in VariableManager get_vars() 9704 1726776648.32515: done with get_vars() 9704 1726776648.32516: filtering new block on tags 9704 1726776648.32540: done filtering new block on tags 9704 1726776648.32542: in VariableManager get_vars() 9704 1726776648.32550: done with get_vars() 9704 1726776648.32552: filtering new block on tags 9704 1726776648.32573: done filtering new block on tags 9704 1726776648.32574: in VariableManager get_vars() 9704 1726776648.32584: done with get_vars() 9704 1726776648.32585: filtering new block on tags 9704 1726776648.32673: done filtering new block on tags 9704 1726776648.32675: in VariableManager get_vars() 9704 1726776648.32684: done with get_vars() 9704 1726776648.32684: filtering new block on tags 9704 1726776648.32696: done filtering new block on tags 9704 1726776648.32697: done iterating over new_blocks loaded from include file 9704 1726776648.32698: extending task lists for all hosts with included blocks 9704 1726776648.32825: done extending task lists 9704 1726776648.32826: done processing included files 9704 1726776648.32827: results queue empty 9704 1726776648.32827: checking for any_errors_fatal 9704 1726776648.32831: done checking for any_errors_fatal 9704 1726776648.32832: checking for max_fail_percentage 9704 1726776648.32832: done checking for max_fail_percentage 9704 1726776648.32833: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.32833: done checking to see if all hosts have failed 9704 1726776648.32834: getting the remaining hosts for this loop 9704 1726776648.32834: done getting the remaining hosts for this loop 9704 1726776648.32836: getting the next task for host managed_node3 9704 1726776648.32838: done getting next task for host managed_node3 9704 1726776648.32840: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values 9704 1726776648.32841: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.32847: getting variables 9704 1726776648.32848: in VariableManager get_vars() 9704 1726776648.32856: Calling all_inventory to load vars for managed_node3 9704 1726776648.32857: Calling groups_inventory to load vars for managed_node3 9704 1726776648.32858: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.32861: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.32863: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.32864: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.32960: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.33072: done with get_vars() 9704 1726776648.33078: done getting variables 9704 1726776648.33122: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:2 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.040) 0:00:01.260 **** 9704 1726776648.33145: entering _queue_task() for managed_node3/fail 9704 1726776648.33146: Creating lock for fail 9704 1726776648.33324: worker is 1 (out of 1 available) 9704 1726776648.33339: exiting _queue_task() for managed_node3/fail 9704 1726776648.33350: done queuing things up, now waiting for results queue to drain 9704 1726776648.33352: waiting for pending results... 9762 1726776648.33450: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values 9762 1726776648.33555: in run() - task 120fa90a-8a95-fee7-0cea-000000000047 9762 1726776648.33570: variable 'ansible_search_path' from source: unknown 9762 1726776648.33575: variable 'ansible_search_path' from source: unknown 9762 1726776648.33603: calling self._execute() 9762 1726776648.33660: variable 'ansible_host' from source: host vars for 'managed_node3' 9762 1726776648.33668: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9762 1726776648.33677: variable 'omit' from source: magic vars 9762 1726776648.33997: variable 'kernel_settings_sysctl' from source: include_vars 9762 1726776648.34014: variable '__kernel_settings_state_empty' from source: role '' all vars 9762 1726776648.34022: Evaluated conditional (kernel_settings_sysctl != __kernel_settings_state_empty): True 9762 1726776648.34230: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9762 1726776648.35898: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9762 1726776648.35952: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9762 1726776648.35980: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9762 1726776648.36005: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9762 1726776648.36032: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9762 1726776648.36086: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9762 1726776648.36110: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9762 1726776648.36133: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9762 1726776648.36161: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9762 1726776648.36172: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9762 1726776648.36206: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9762 1726776648.36222: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9762 1726776648.36239: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9762 1726776648.36262: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9762 1726776648.36270: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9762 1726776648.36296: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9762 1726776648.36311: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9762 1726776648.36325: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9762 1726776648.36360: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9762 1726776648.36372: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9762 1726776648.36551: variable 'kernel_settings_sysctl' from source: include_vars 9762 1726776648.36606: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 9762 1726776648.36727: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 9762 1726776648.36755: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 9762 1726776648.36778: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 9762 1726776648.36801: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 9762 1726776648.36836: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9762 1726776648.36853: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9762 1726776648.36870: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9762 1726776648.36887: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9762 1726776648.36922: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9762 1726776648.36940: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9762 1726776648.36957: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9762 1726776648.36975: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9762 1726776648.36996: Evaluated conditional ((kernel_settings_sysctl | selectattr("value", "defined") | selectattr("value", "sameas", true) | list | length > 0) or (kernel_settings_sysctl | selectattr("value", "defined") | selectattr("value", "sameas", false) | list | length > 0)): False 9762 1726776648.37000: when evaluation is False, skipping this task 9762 1726776648.37004: _execute() done 9762 1726776648.37010: dumping result to json 9762 1726776648.37013: done dumping result, returning 9762 1726776648.37020: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values [120fa90a-8a95-fee7-0cea-000000000047] 9762 1726776648.37025: sending task result for task 120fa90a-8a95-fee7-0cea-000000000047 9762 1726776648.37053: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000047 9762 1726776648.37056: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "(kernel_settings_sysctl | selectattr(\"value\", \"defined\") | selectattr(\"value\", \"sameas\", true) | list | length > 0) or (kernel_settings_sysctl | selectattr(\"value\", \"defined\") | selectattr(\"value\", \"sameas\", false) | list | length > 0)", "skip_reason": "Conditional result was False" } 9704 1726776648.37177: no more pending results, returning what we have 9704 1726776648.37180: results queue empty 9704 1726776648.37180: checking for any_errors_fatal 9704 1726776648.37182: done checking for any_errors_fatal 9704 1726776648.37182: checking for max_fail_percentage 9704 1726776648.37184: done checking for max_fail_percentage 9704 1726776648.37184: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.37185: done checking to see if all hosts have failed 9704 1726776648.37185: getting the remaining hosts for this loop 9704 1726776648.37187: done getting the remaining hosts for this loop 9704 1726776648.37190: getting the next task for host managed_node3 9704 1726776648.37195: done getting next task for host managed_node3 9704 1726776648.37199: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables 9704 1726776648.37201: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.37213: getting variables 9704 1726776648.37214: in VariableManager get_vars() 9704 1726776648.37246: Calling all_inventory to load vars for managed_node3 9704 1726776648.37248: Calling groups_inventory to load vars for managed_node3 9704 1726776648.37250: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.37258: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.37260: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.37262: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.37377: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.37490: done with get_vars() 9704 1726776648.37498: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Set version specific variables] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:9 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.044) 0:00:01.304 **** 9704 1726776648.37563: entering _queue_task() for managed_node3/include_tasks 9704 1726776648.37564: Creating lock for include_tasks 9704 1726776648.37726: worker is 1 (out of 1 available) 9704 1726776648.37739: exiting _queue_task() for managed_node3/include_tasks 9704 1726776648.37750: done queuing things up, now waiting for results queue to drain 9704 1726776648.37751: waiting for pending results... 9763 1726776648.37848: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables 9763 1726776648.37949: in run() - task 120fa90a-8a95-fee7-0cea-000000000048 9763 1726776648.37964: variable 'ansible_search_path' from source: unknown 9763 1726776648.37968: variable 'ansible_search_path' from source: unknown 9763 1726776648.37996: calling self._execute() 9763 1726776648.38053: variable 'ansible_host' from source: host vars for 'managed_node3' 9763 1726776648.38061: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9763 1726776648.38070: variable 'omit' from source: magic vars 9763 1726776648.38142: _execute() done 9763 1726776648.38148: dumping result to json 9763 1726776648.38152: done dumping result, returning 9763 1726776648.38159: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables [120fa90a-8a95-fee7-0cea-000000000048] 9763 1726776648.38167: sending task result for task 120fa90a-8a95-fee7-0cea-000000000048 9763 1726776648.38188: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000048 9763 1726776648.38192: WORKER PROCESS EXITING 9704 1726776648.38286: no more pending results, returning what we have 9704 1726776648.38290: in VariableManager get_vars() 9704 1726776648.38322: Calling all_inventory to load vars for managed_node3 9704 1726776648.38325: Calling groups_inventory to load vars for managed_node3 9704 1726776648.38327: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.38336: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.38338: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.38340: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.38469: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.38577: done with get_vars() 9704 1726776648.38584: variable 'ansible_search_path' from source: unknown 9704 1726776648.38584: variable 'ansible_search_path' from source: unknown 9704 1726776648.38607: we have included files to process 9704 1726776648.38608: generating all_blocks data 9704 1726776648.38609: done generating all_blocks data 9704 1726776648.38613: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml 9704 1726776648.38614: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml 9704 1726776648.38615: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml for managed_node3 9704 1726776648.39127: done processing included file 9704 1726776648.39130: iterating over new_blocks loaded from include file 9704 1726776648.39131: in VariableManager get_vars() 9704 1726776648.39144: done with get_vars() 9704 1726776648.39145: filtering new block on tags 9704 1726776648.39161: done filtering new block on tags 9704 1726776648.39163: in VariableManager get_vars() 9704 1726776648.39175: done with get_vars() 9704 1726776648.39176: filtering new block on tags 9704 1726776648.39197: done filtering new block on tags 9704 1726776648.39198: in VariableManager get_vars() 9704 1726776648.39210: done with get_vars() 9704 1726776648.39212: filtering new block on tags 9704 1726776648.39237: done filtering new block on tags 9704 1726776648.39353: in VariableManager get_vars() 9704 1726776648.39371: done with get_vars() 9704 1726776648.39372: filtering new block on tags 9704 1726776648.39387: done filtering new block on tags 9704 1726776648.39388: done iterating over new_blocks loaded from include file 9704 1726776648.39389: extending task lists for all hosts with included blocks 9704 1726776648.39479: done extending task lists 9704 1726776648.39480: done processing included files 9704 1726776648.39480: results queue empty 9704 1726776648.39480: checking for any_errors_fatal 9704 1726776648.39482: done checking for any_errors_fatal 9704 1726776648.39482: checking for max_fail_percentage 9704 1726776648.39483: done checking for max_fail_percentage 9704 1726776648.39483: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.39484: done checking to see if all hosts have failed 9704 1726776648.39484: getting the remaining hosts for this loop 9704 1726776648.39485: done getting the remaining hosts for this loop 9704 1726776648.39486: getting the next task for host managed_node3 9704 1726776648.39489: done getting next task for host managed_node3 9704 1726776648.39490: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role 9704 1726776648.39492: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.39497: getting variables 9704 1726776648.39498: in VariableManager get_vars() 9704 1726776648.39505: Calling all_inventory to load vars for managed_node3 9704 1726776648.39506: Calling groups_inventory to load vars for managed_node3 9704 1726776648.39508: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.39511: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.39512: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.39514: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.39589: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.39694: done with get_vars() 9704 1726776648.39700: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:2 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.021) 0:00:01.326 **** 9704 1726776648.39745: entering _queue_task() for managed_node3/setup 9704 1726776648.39882: worker is 1 (out of 1 available) 9704 1726776648.39894: exiting _queue_task() for managed_node3/setup 9704 1726776648.39903: done queuing things up, now waiting for results queue to drain 9704 1726776648.39905: waiting for pending results... 9764 1726776648.39999: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role 9764 1726776648.40105: in run() - task 120fa90a-8a95-fee7-0cea-0000000000bb 9764 1726776648.40121: variable 'ansible_search_path' from source: unknown 9764 1726776648.40125: variable 'ansible_search_path' from source: unknown 9764 1726776648.40152: calling self._execute() 9764 1726776648.40203: variable 'ansible_host' from source: host vars for 'managed_node3' 9764 1726776648.40212: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9764 1726776648.40217: variable 'omit' from source: magic vars 9764 1726776648.40570: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9764 1726776648.42071: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9764 1726776648.42119: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9764 1726776648.42155: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9764 1726776648.42181: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9764 1726776648.42202: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9764 1726776648.42258: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9764 1726776648.42279: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9764 1726776648.42297: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9764 1726776648.42330: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9764 1726776648.42343: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9764 1726776648.42378: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9764 1726776648.42395: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9764 1726776648.42417: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9764 1726776648.42445: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9764 1726776648.42457: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9764 1726776648.42571: variable '__kernel_settings_required_facts' from source: role '' all vars 9764 1726776648.42581: variable 'ansible_facts' from source: unknown 9764 1726776648.42639: Evaluated conditional (__kernel_settings_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 9764 1726776648.42644: when evaluation is False, skipping this task 9764 1726776648.42647: _execute() done 9764 1726776648.42651: dumping result to json 9764 1726776648.42654: done dumping result, returning 9764 1726776648.42659: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role [120fa90a-8a95-fee7-0cea-0000000000bb] 9764 1726776648.42663: sending task result for task 120fa90a-8a95-fee7-0cea-0000000000bb 9764 1726776648.42681: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000000bb 9764 1726776648.42683: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } 9704 1726776648.42813: no more pending results, returning what we have 9704 1726776648.42817: results queue empty 9704 1726776648.42817: checking for any_errors_fatal 9704 1726776648.42819: done checking for any_errors_fatal 9704 1726776648.42820: checking for max_fail_percentage 9704 1726776648.42821: done checking for max_fail_percentage 9704 1726776648.42821: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.42822: done checking to see if all hosts have failed 9704 1726776648.42823: getting the remaining hosts for this loop 9704 1726776648.42824: done getting the remaining hosts for this loop 9704 1726776648.42827: getting the next task for host managed_node3 9704 1726776648.42836: done getting next task for host managed_node3 9704 1726776648.42839: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree 9704 1726776648.42842: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.42855: getting variables 9704 1726776648.42856: in VariableManager get_vars() 9704 1726776648.42884: Calling all_inventory to load vars for managed_node3 9704 1726776648.42886: Calling groups_inventory to load vars for managed_node3 9704 1726776648.42888: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.42895: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.42897: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.42899: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.43024: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.43140: done with get_vars() 9704 1726776648.43146: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Check if system is ostree] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:10 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.034) 0:00:01.360 **** 9704 1726776648.43204: entering _queue_task() for managed_node3/stat 9704 1726776648.43346: worker is 1 (out of 1 available) 9704 1726776648.43359: exiting _queue_task() for managed_node3/stat 9704 1726776648.43369: done queuing things up, now waiting for results queue to drain 9704 1726776648.43371: waiting for pending results... 9765 1726776648.43466: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree 9765 1726776648.43575: in run() - task 120fa90a-8a95-fee7-0cea-0000000000bd 9765 1726776648.43590: variable 'ansible_search_path' from source: unknown 9765 1726776648.43594: variable 'ansible_search_path' from source: unknown 9765 1726776648.43622: calling self._execute() 9765 1726776648.43673: variable 'ansible_host' from source: host vars for 'managed_node3' 9765 1726776648.43682: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9765 1726776648.43690: variable 'omit' from source: magic vars 9765 1726776648.43995: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 9765 1726776648.44159: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 9765 1726776648.44192: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 9765 1726776648.44220: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 9765 1726776648.44249: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 9765 1726776648.44305: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9765 1726776648.44327: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9765 1726776648.44348: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9765 1726776648.44366: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9765 1726776648.44456: Evaluated conditional (not __kernel_settings_is_ostree is defined): True 9765 1726776648.44464: variable 'omit' from source: magic vars 9765 1726776648.44502: variable 'omit' from source: magic vars 9765 1726776648.44527: variable 'omit' from source: magic vars 9765 1726776648.44548: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9765 1726776648.44569: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9765 1726776648.44585: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9765 1726776648.44598: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9765 1726776648.44610: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9765 1726776648.44635: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9765 1726776648.44640: variable 'ansible_host' from source: host vars for 'managed_node3' 9765 1726776648.44644: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9765 1726776648.44705: Set connection var ansible_module_compression to ZIP_DEFLATED 9765 1726776648.44716: Set connection var ansible_shell_type to sh 9765 1726776648.44724: Set connection var ansible_timeout to 10 9765 1726776648.44730: Set connection var ansible_pipelining to False 9765 1726776648.44738: Set connection var ansible_shell_executable to /bin/sh 9765 1726776648.44741: Set connection var ansible_connection to ssh 9765 1726776648.44755: variable 'ansible_shell_executable' from source: unknown 9765 1726776648.44759: variable 'ansible_connection' from source: unknown 9765 1726776648.44762: variable 'ansible_module_compression' from source: unknown 9765 1726776648.44766: variable 'ansible_shell_type' from source: unknown 9765 1726776648.44769: variable 'ansible_shell_executable' from source: unknown 9765 1726776648.44772: variable 'ansible_host' from source: host vars for 'managed_node3' 9765 1726776648.44777: variable 'ansible_pipelining' from source: unknown 9765 1726776648.44780: variable 'ansible_timeout' from source: unknown 9765 1726776648.44784: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9765 1726776648.44874: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 9765 1726776648.44885: variable 'omit' from source: magic vars 9765 1726776648.44890: starting attempt loop 9765 1726776648.44894: running the handler 9765 1726776648.44903: _low_level_execute_command(): starting 9765 1726776648.44911: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9765 1726776648.47211: stdout chunk (state=2): >>>/root <<< 9765 1726776648.47330: stderr chunk (state=3): >>><<< 9765 1726776648.47337: stdout chunk (state=3): >>><<< 9765 1726776648.47353: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9765 1726776648.47366: _low_level_execute_command(): starting 9765 1726776648.47371: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084 `" && echo ansible-tmp-1726776648.4736056-9765-24254485172084="` echo /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084 `" ) && sleep 0' 9765 1726776648.49805: stdout chunk (state=2): >>>ansible-tmp-1726776648.4736056-9765-24254485172084=/root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084 <<< 9765 1726776648.49931: stderr chunk (state=3): >>><<< 9765 1726776648.49938: stdout chunk (state=3): >>><<< 9765 1726776648.49952: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776648.4736056-9765-24254485172084=/root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084 , stderr= 9765 1726776648.49989: variable 'ansible_module_compression' from source: unknown 9765 1726776648.50037: ANSIBALLZ: Using lock for stat 9765 1726776648.50041: ANSIBALLZ: Acquiring lock 9765 1726776648.50045: ANSIBALLZ: Lock acquired: 140007686721120 9765 1726776648.50049: ANSIBALLZ: Creating module 9765 1726776648.59314: ANSIBALLZ: Writing module into payload 9765 1726776648.59399: ANSIBALLZ: Writing module 9765 1726776648.59419: ANSIBALLZ: Renaming module 9765 1726776648.59426: ANSIBALLZ: Done creating module 9765 1726776648.59443: variable 'ansible_facts' from source: unknown 9765 1726776648.59499: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/AnsiballZ_stat.py 9765 1726776648.59596: Sending initial data 9765 1726776648.59603: Sent initial data (150 bytes) 9765 1726776648.62141: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp84m6umob /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/AnsiballZ_stat.py <<< 9765 1726776648.63119: stderr chunk (state=3): >>><<< 9765 1726776648.63129: stdout chunk (state=3): >>><<< 9765 1726776648.63150: done transferring module to remote 9765 1726776648.63161: _low_level_execute_command(): starting 9765 1726776648.63166: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/ /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/AnsiballZ_stat.py && sleep 0' 9765 1726776648.65543: stderr chunk (state=2): >>><<< 9765 1726776648.65551: stdout chunk (state=2): >>><<< 9765 1726776648.65565: _low_level_execute_command() done: rc=0, stdout=, stderr= 9765 1726776648.65570: _low_level_execute_command(): starting 9765 1726776648.65575: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/AnsiballZ_stat.py && sleep 0' 9765 1726776648.80502: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/run/ostree-booted", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 9765 1726776648.81588: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9765 1726776648.81598: stdout chunk (state=3): >>><<< 9765 1726776648.81612: stderr chunk (state=3): >>><<< 9765 1726776648.81626: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/run/ostree-booted", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 9765 1726776648.81654: done with _execute_module (stat, {'path': '/run/ostree-booted', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9765 1726776648.81663: _low_level_execute_command(): starting 9765 1726776648.81667: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776648.4736056-9765-24254485172084/ > /dev/null 2>&1 && sleep 0' 9765 1726776648.84052: stderr chunk (state=2): >>><<< 9765 1726776648.84060: stdout chunk (state=2): >>><<< 9765 1726776648.84074: _low_level_execute_command() done: rc=0, stdout=, stderr= 9765 1726776648.84081: handler run complete 9765 1726776648.84096: attempt loop complete, returning result 9765 1726776648.84100: _execute() done 9765 1726776648.84103: dumping result to json 9765 1726776648.84110: done dumping result, returning 9765 1726776648.84118: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree [120fa90a-8a95-fee7-0cea-0000000000bd] 9765 1726776648.84124: sending task result for task 120fa90a-8a95-fee7-0cea-0000000000bd 9765 1726776648.84152: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000000bd 9765 1726776648.84156: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "stat": { "exists": false } } 9704 1726776648.84291: no more pending results, returning what we have 9704 1726776648.84294: results queue empty 9704 1726776648.84295: checking for any_errors_fatal 9704 1726776648.84299: done checking for any_errors_fatal 9704 1726776648.84299: checking for max_fail_percentage 9704 1726776648.84301: done checking for max_fail_percentage 9704 1726776648.84301: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.84302: done checking to see if all hosts have failed 9704 1726776648.84302: getting the remaining hosts for this loop 9704 1726776648.84304: done getting the remaining hosts for this loop 9704 1726776648.84307: getting the next task for host managed_node3 9704 1726776648.84312: done getting next task for host managed_node3 9704 1726776648.84315: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree 9704 1726776648.84319: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.84327: getting variables 9704 1726776648.84330: in VariableManager get_vars() 9704 1726776648.84360: Calling all_inventory to load vars for managed_node3 9704 1726776648.84363: Calling groups_inventory to load vars for managed_node3 9704 1726776648.84364: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.84373: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.84375: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.84377: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.84510: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.84631: done with get_vars() 9704 1726776648.84640: done getting variables 9704 1726776648.84680: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:15 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.414) 0:00:01.775 **** 9704 1726776648.84702: entering _queue_task() for managed_node3/set_fact 9704 1726776648.84863: worker is 1 (out of 1 available) 9704 1726776648.84878: exiting _queue_task() for managed_node3/set_fact 9704 1726776648.84890: done queuing things up, now waiting for results queue to drain 9704 1726776648.84892: waiting for pending results... 9776 1726776648.84993: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree 9776 1726776648.85112: in run() - task 120fa90a-8a95-fee7-0cea-0000000000be 9776 1726776648.85127: variable 'ansible_search_path' from source: unknown 9776 1726776648.85134: variable 'ansible_search_path' from source: unknown 9776 1726776648.85162: calling self._execute() 9776 1726776648.85220: variable 'ansible_host' from source: host vars for 'managed_node3' 9776 1726776648.85231: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9776 1726776648.85240: variable 'omit' from source: magic vars 9776 1726776648.85568: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 9776 1726776648.85794: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 9776 1726776648.85830: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 9776 1726776648.85856: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 9776 1726776648.85882: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 9776 1726776648.85945: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9776 1726776648.85965: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9776 1726776648.85984: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9776 1726776648.86004: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9776 1726776648.86089: Evaluated conditional (not __kernel_settings_is_ostree is defined): True 9776 1726776648.86098: variable 'omit' from source: magic vars 9776 1726776648.86140: variable 'omit' from source: magic vars 9776 1726776648.86220: variable '__ostree_booted_stat' from source: set_fact 9776 1726776648.86259: variable 'omit' from source: magic vars 9776 1726776648.86278: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9776 1726776648.86297: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9776 1726776648.86316: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9776 1726776648.86333: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9776 1726776648.86343: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9776 1726776648.86366: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9776 1726776648.86370: variable 'ansible_host' from source: host vars for 'managed_node3' 9776 1726776648.86374: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9776 1726776648.86441: Set connection var ansible_module_compression to ZIP_DEFLATED 9776 1726776648.86449: Set connection var ansible_shell_type to sh 9776 1726776648.86458: Set connection var ansible_timeout to 10 9776 1726776648.86464: Set connection var ansible_pipelining to False 9776 1726776648.86471: Set connection var ansible_shell_executable to /bin/sh 9776 1726776648.86474: Set connection var ansible_connection to ssh 9776 1726776648.86489: variable 'ansible_shell_executable' from source: unknown 9776 1726776648.86492: variable 'ansible_connection' from source: unknown 9776 1726776648.86495: variable 'ansible_module_compression' from source: unknown 9776 1726776648.86499: variable 'ansible_shell_type' from source: unknown 9776 1726776648.86502: variable 'ansible_shell_executable' from source: unknown 9776 1726776648.86505: variable 'ansible_host' from source: host vars for 'managed_node3' 9776 1726776648.86511: variable 'ansible_pipelining' from source: unknown 9776 1726776648.86515: variable 'ansible_timeout' from source: unknown 9776 1726776648.86519: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9776 1726776648.86583: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9776 1726776648.86594: variable 'omit' from source: magic vars 9776 1726776648.86600: starting attempt loop 9776 1726776648.86604: running the handler 9776 1726776648.86614: handler run complete 9776 1726776648.86622: attempt loop complete, returning result 9776 1726776648.86625: _execute() done 9776 1726776648.86628: dumping result to json 9776 1726776648.86634: done dumping result, returning 9776 1726776648.86640: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree [120fa90a-8a95-fee7-0cea-0000000000be] 9776 1726776648.86645: sending task result for task 120fa90a-8a95-fee7-0cea-0000000000be 9776 1726776648.86660: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000000be 9776 1726776648.86662: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_is_ostree": false }, "changed": false } 9704 1726776648.86885: no more pending results, returning what we have 9704 1726776648.86887: results queue empty 9704 1726776648.86887: checking for any_errors_fatal 9704 1726776648.86894: done checking for any_errors_fatal 9704 1726776648.86894: checking for max_fail_percentage 9704 1726776648.86895: done checking for max_fail_percentage 9704 1726776648.86896: checking to see if all hosts have failed and the running result is not ok 9704 1726776648.86896: done checking to see if all hosts have failed 9704 1726776648.86897: getting the remaining hosts for this loop 9704 1726776648.86897: done getting the remaining hosts for this loop 9704 1726776648.86900: getting the next task for host managed_node3 9704 1726776648.86905: done getting next task for host managed_node3 9704 1726776648.86908: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin 9704 1726776648.86910: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776648.86917: getting variables 9704 1726776648.86918: in VariableManager get_vars() 9704 1726776648.86943: Calling all_inventory to load vars for managed_node3 9704 1726776648.86945: Calling groups_inventory to load vars for managed_node3 9704 1726776648.86946: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776648.86951: Calling all_plugins_play to load vars for managed_node3 9704 1726776648.86953: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776648.86954: Calling groups_plugins_play to load vars for managed_node3 9704 1726776648.87083: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776648.87195: done with get_vars() 9704 1726776648.87202: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:22 Thursday 19 September 2024 16:10:48 -0400 (0:00:00.025) 0:00:01.801 **** 9704 1726776648.87265: entering _queue_task() for managed_node3/stat 9704 1726776648.87406: worker is 1 (out of 1 available) 9704 1726776648.87418: exiting _queue_task() for managed_node3/stat 9704 1726776648.87430: done queuing things up, now waiting for results queue to drain 9704 1726776648.87432: waiting for pending results... 9777 1726776648.87528: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin 9777 1726776648.87643: in run() - task 120fa90a-8a95-fee7-0cea-0000000000c0 9777 1726776648.87658: variable 'ansible_search_path' from source: unknown 9777 1726776648.87662: variable 'ansible_search_path' from source: unknown 9777 1726776648.87688: calling self._execute() 9777 1726776648.87743: variable 'ansible_host' from source: host vars for 'managed_node3' 9777 1726776648.87751: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9777 1726776648.87760: variable 'omit' from source: magic vars 9777 1726776648.88075: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 9777 1726776648.88244: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 9777 1726776648.88275: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 9777 1726776648.88301: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 9777 1726776648.88332: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 9777 1726776648.88386: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9777 1726776648.88409: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9777 1726776648.88432: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9777 1726776648.88451: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9777 1726776648.88533: Evaluated conditional (not __kernel_settings_is_transactional is defined): True 9777 1726776648.88541: variable 'omit' from source: magic vars 9777 1726776648.88580: variable 'omit' from source: magic vars 9777 1726776648.88601: variable 'omit' from source: magic vars 9777 1726776648.88622: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9777 1726776648.88645: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9777 1726776648.88660: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9777 1726776648.88673: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9777 1726776648.88682: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9777 1726776648.88703: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9777 1726776648.88710: variable 'ansible_host' from source: host vars for 'managed_node3' 9777 1726776648.88715: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9777 1726776648.88779: Set connection var ansible_module_compression to ZIP_DEFLATED 9777 1726776648.88787: Set connection var ansible_shell_type to sh 9777 1726776648.88795: Set connection var ansible_timeout to 10 9777 1726776648.88801: Set connection var ansible_pipelining to False 9777 1726776648.88809: Set connection var ansible_shell_executable to /bin/sh 9777 1726776648.88813: Set connection var ansible_connection to ssh 9777 1726776648.88828: variable 'ansible_shell_executable' from source: unknown 9777 1726776648.88834: variable 'ansible_connection' from source: unknown 9777 1726776648.88836: variable 'ansible_module_compression' from source: unknown 9777 1726776648.88838: variable 'ansible_shell_type' from source: unknown 9777 1726776648.88840: variable 'ansible_shell_executable' from source: unknown 9777 1726776648.88841: variable 'ansible_host' from source: host vars for 'managed_node3' 9777 1726776648.88843: variable 'ansible_pipelining' from source: unknown 9777 1726776648.88845: variable 'ansible_timeout' from source: unknown 9777 1726776648.88847: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9777 1726776648.88949: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 9777 1726776648.88959: variable 'omit' from source: magic vars 9777 1726776648.88965: starting attempt loop 9777 1726776648.88969: running the handler 9777 1726776648.88978: _low_level_execute_command(): starting 9777 1726776648.88983: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9777 1726776648.91271: stdout chunk (state=2): >>>/root <<< 9777 1726776648.91390: stderr chunk (state=3): >>><<< 9777 1726776648.91396: stdout chunk (state=3): >>><<< 9777 1726776648.91414: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9777 1726776648.91427: _low_level_execute_command(): starting 9777 1726776648.91434: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509 `" && echo ansible-tmp-1726776648.9142194-9777-8091946225509="` echo /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509 `" ) && sleep 0' 9777 1726776648.93833: stdout chunk (state=2): >>>ansible-tmp-1726776648.9142194-9777-8091946225509=/root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509 <<< 9777 1726776648.93960: stderr chunk (state=3): >>><<< 9777 1726776648.93966: stdout chunk (state=3): >>><<< 9777 1726776648.93980: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776648.9142194-9777-8091946225509=/root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509 , stderr= 9777 1726776648.94015: variable 'ansible_module_compression' from source: unknown 9777 1726776648.94060: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 9777 1726776648.94088: variable 'ansible_facts' from source: unknown 9777 1726776648.94158: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/AnsiballZ_stat.py 9777 1726776648.94284: Sending initial data 9777 1726776648.94291: Sent initial data (149 bytes) 9777 1726776648.96777: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpx87n1f7h /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/AnsiballZ_stat.py <<< 9777 1726776648.97752: stderr chunk (state=3): >>><<< 9777 1726776648.97760: stdout chunk (state=3): >>><<< 9777 1726776648.97778: done transferring module to remote 9777 1726776648.97790: _low_level_execute_command(): starting 9777 1726776648.97796: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/ /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/AnsiballZ_stat.py && sleep 0' 9777 1726776649.00149: stderr chunk (state=2): >>><<< 9777 1726776649.00156: stdout chunk (state=2): >>><<< 9777 1726776649.00169: _low_level_execute_command() done: rc=0, stdout=, stderr= 9777 1726776649.00174: _low_level_execute_command(): starting 9777 1726776649.00180: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/AnsiballZ_stat.py && sleep 0' 9777 1726776649.15371: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/sbin/transactional-update", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 9777 1726776649.16414: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9777 1726776649.16466: stderr chunk (state=3): >>><<< 9777 1726776649.16473: stdout chunk (state=3): >>><<< 9777 1726776649.16489: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/sbin/transactional-update", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 9777 1726776649.16518: done with _execute_module (stat, {'path': '/sbin/transactional-update', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9777 1726776649.16531: _low_level_execute_command(): starting 9777 1726776649.16537: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776648.9142194-9777-8091946225509/ > /dev/null 2>&1 && sleep 0' 9777 1726776649.18942: stderr chunk (state=2): >>><<< 9777 1726776649.18952: stdout chunk (state=2): >>><<< 9777 1726776649.18969: _low_level_execute_command() done: rc=0, stdout=, stderr= 9777 1726776649.18976: handler run complete 9777 1726776649.18991: attempt loop complete, returning result 9777 1726776649.18994: _execute() done 9777 1726776649.18998: dumping result to json 9777 1726776649.19002: done dumping result, returning 9777 1726776649.19010: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin [120fa90a-8a95-fee7-0cea-0000000000c0] 9777 1726776649.19017: sending task result for task 120fa90a-8a95-fee7-0cea-0000000000c0 9777 1726776649.19049: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000000c0 9777 1726776649.19052: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "stat": { "exists": false } } 9704 1726776649.19188: no more pending results, returning what we have 9704 1726776649.19191: results queue empty 9704 1726776649.19191: checking for any_errors_fatal 9704 1726776649.19197: done checking for any_errors_fatal 9704 1726776649.19197: checking for max_fail_percentage 9704 1726776649.19198: done checking for max_fail_percentage 9704 1726776649.19199: checking to see if all hosts have failed and the running result is not ok 9704 1726776649.19200: done checking to see if all hosts have failed 9704 1726776649.19200: getting the remaining hosts for this loop 9704 1726776649.19201: done getting the remaining hosts for this loop 9704 1726776649.19204: getting the next task for host managed_node3 9704 1726776649.19212: done getting next task for host managed_node3 9704 1726776649.19216: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists 9704 1726776649.19218: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776649.19230: getting variables 9704 1726776649.19231: in VariableManager get_vars() 9704 1726776649.19262: Calling all_inventory to load vars for managed_node3 9704 1726776649.19265: Calling groups_inventory to load vars for managed_node3 9704 1726776649.19266: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776649.19274: Calling all_plugins_play to load vars for managed_node3 9704 1726776649.19276: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776649.19278: Calling groups_plugins_play to load vars for managed_node3 9704 1726776649.19392: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776649.19511: done with get_vars() 9704 1726776649.19520: done getting variables 9704 1726776649.19562: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:27 Thursday 19 September 2024 16:10:49 -0400 (0:00:00.323) 0:00:02.124 **** 9704 1726776649.19586: entering _queue_task() for managed_node3/set_fact 9704 1726776649.19754: worker is 1 (out of 1 available) 9704 1726776649.19768: exiting _queue_task() for managed_node3/set_fact 9704 1726776649.19779: done queuing things up, now waiting for results queue to drain 9704 1726776649.19781: waiting for pending results... 9785 1726776649.19880: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists 9785 1726776649.19992: in run() - task 120fa90a-8a95-fee7-0cea-0000000000c1 9785 1726776649.20009: variable 'ansible_search_path' from source: unknown 9785 1726776649.20014: variable 'ansible_search_path' from source: unknown 9785 1726776649.20044: calling self._execute() 9785 1726776649.20095: variable 'ansible_host' from source: host vars for 'managed_node3' 9785 1726776649.20103: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9785 1726776649.20111: variable 'omit' from source: magic vars 9785 1726776649.20436: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 9785 1726776649.20661: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 9785 1726776649.20695: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 9785 1726776649.20721: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 9785 1726776649.20750: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 9785 1726776649.20811: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9785 1726776649.20834: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9785 1726776649.20852: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9785 1726776649.20870: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9785 1726776649.20960: Evaluated conditional (not __kernel_settings_is_transactional is defined): True 9785 1726776649.20969: variable 'omit' from source: magic vars 9785 1726776649.21012: variable 'omit' from source: magic vars 9785 1726776649.21095: variable '__transactional_update_stat' from source: set_fact 9785 1726776649.21135: variable 'omit' from source: magic vars 9785 1726776649.21155: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9785 1726776649.21175: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9785 1726776649.21191: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9785 1726776649.21205: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9785 1726776649.21214: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9785 1726776649.21240: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9785 1726776649.21246: variable 'ansible_host' from source: host vars for 'managed_node3' 9785 1726776649.21251: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9785 1726776649.21313: Set connection var ansible_module_compression to ZIP_DEFLATED 9785 1726776649.21321: Set connection var ansible_shell_type to sh 9785 1726776649.21331: Set connection var ansible_timeout to 10 9785 1726776649.21337: Set connection var ansible_pipelining to False 9785 1726776649.21344: Set connection var ansible_shell_executable to /bin/sh 9785 1726776649.21347: Set connection var ansible_connection to ssh 9785 1726776649.21363: variable 'ansible_shell_executable' from source: unknown 9785 1726776649.21368: variable 'ansible_connection' from source: unknown 9785 1726776649.21371: variable 'ansible_module_compression' from source: unknown 9785 1726776649.21374: variable 'ansible_shell_type' from source: unknown 9785 1726776649.21378: variable 'ansible_shell_executable' from source: unknown 9785 1726776649.21381: variable 'ansible_host' from source: host vars for 'managed_node3' 9785 1726776649.21385: variable 'ansible_pipelining' from source: unknown 9785 1726776649.21388: variable 'ansible_timeout' from source: unknown 9785 1726776649.21392: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9785 1726776649.21457: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9785 1726776649.21467: variable 'omit' from source: magic vars 9785 1726776649.21472: starting attempt loop 9785 1726776649.21474: running the handler 9785 1726776649.21481: handler run complete 9785 1726776649.21486: attempt loop complete, returning result 9785 1726776649.21489: _execute() done 9785 1726776649.21490: dumping result to json 9785 1726776649.21492: done dumping result, returning 9785 1726776649.21496: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists [120fa90a-8a95-fee7-0cea-0000000000c1] 9785 1726776649.21500: sending task result for task 120fa90a-8a95-fee7-0cea-0000000000c1 9785 1726776649.21515: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000000c1 9785 1726776649.21517: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_is_transactional": false }, "changed": false } 9704 1726776649.21779: no more pending results, returning what we have 9704 1726776649.21781: results queue empty 9704 1726776649.21782: checking for any_errors_fatal 9704 1726776649.21788: done checking for any_errors_fatal 9704 1726776649.21789: checking for max_fail_percentage 9704 1726776649.21790: done checking for max_fail_percentage 9704 1726776649.21790: checking to see if all hosts have failed and the running result is not ok 9704 1726776649.21790: done checking to see if all hosts have failed 9704 1726776649.21791: getting the remaining hosts for this loop 9704 1726776649.21792: done getting the remaining hosts for this loop 9704 1726776649.21794: getting the next task for host managed_node3 9704 1726776649.21800: done getting next task for host managed_node3 9704 1726776649.21802: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables 9704 1726776649.21804: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776649.21813: getting variables 9704 1726776649.21818: in VariableManager get_vars() 9704 1726776649.21842: Calling all_inventory to load vars for managed_node3 9704 1726776649.21844: Calling groups_inventory to load vars for managed_node3 9704 1726776649.21845: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776649.21850: Calling all_plugins_play to load vars for managed_node3 9704 1726776649.21852: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776649.21853: Calling groups_plugins_play to load vars for managed_node3 9704 1726776649.21987: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776649.22105: done with get_vars() 9704 1726776649.22114: done getting variables 9704 1726776649.22153: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set platform/version specific variables] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:31 Thursday 19 September 2024 16:10:49 -0400 (0:00:00.025) 0:00:02.150 **** 9704 1726776649.22175: entering _queue_task() for managed_node3/include_vars 9704 1726776649.22333: worker is 1 (out of 1 available) 9704 1726776649.22346: exiting _queue_task() for managed_node3/include_vars 9704 1726776649.22357: done queuing things up, now waiting for results queue to drain 9704 1726776649.22359: waiting for pending results... 9786 1726776649.22455: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables 9786 1726776649.22566: in run() - task 120fa90a-8a95-fee7-0cea-0000000000c3 9786 1726776649.22581: variable 'ansible_search_path' from source: unknown 9786 1726776649.22586: variable 'ansible_search_path' from source: unknown 9786 1726776649.22611: calling self._execute() 9786 1726776649.22665: variable 'ansible_host' from source: host vars for 'managed_node3' 9786 1726776649.22673: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9786 1726776649.22681: variable 'omit' from source: magic vars 9786 1726776649.22752: variable 'omit' from source: magic vars 9786 1726776649.22794: variable 'omit' from source: magic vars 9786 1726776649.23046: variable 'ffparams' from source: task vars 9786 1726776649.23138: variable 'ansible_facts' from source: unknown 9786 1726776649.23263: variable 'ansible_facts' from source: unknown 9786 1726776649.23352: variable 'ansible_facts' from source: unknown 9786 1726776649.23440: variable 'ansible_facts' from source: unknown 9786 1726776649.23516: variable 'role_path' from source: magic vars 9786 1726776649.23630: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 9786 1726776649.23786: Loaded config def from plugin (lookup/first_found) 9786 1726776649.23795: Loading LookupModule 'first_found' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/first_found.py 9786 1726776649.23823: variable 'ansible_search_path' from source: unknown 9786 1726776649.23843: variable 'ansible_search_path' from source: unknown 9786 1726776649.23852: variable 'ansible_search_path' from source: unknown 9786 1726776649.23859: variable 'ansible_search_path' from source: unknown 9786 1726776649.23866: variable 'ansible_search_path' from source: unknown 9786 1726776649.23880: variable 'omit' from source: magic vars 9786 1726776649.23895: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9786 1726776649.23911: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9786 1726776649.23924: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9786 1726776649.23945: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9786 1726776649.23955: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9786 1726776649.23974: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9786 1726776649.23977: variable 'ansible_host' from source: host vars for 'managed_node3' 9786 1726776649.23979: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9786 1726776649.24037: Set connection var ansible_module_compression to ZIP_DEFLATED 9786 1726776649.24043: Set connection var ansible_shell_type to sh 9786 1726776649.24048: Set connection var ansible_timeout to 10 9786 1726776649.24051: Set connection var ansible_pipelining to False 9786 1726776649.24055: Set connection var ansible_shell_executable to /bin/sh 9786 1726776649.24057: Set connection var ansible_connection to ssh 9786 1726776649.24072: variable 'ansible_shell_executable' from source: unknown 9786 1726776649.24075: variable 'ansible_connection' from source: unknown 9786 1726776649.24077: variable 'ansible_module_compression' from source: unknown 9786 1726776649.24079: variable 'ansible_shell_type' from source: unknown 9786 1726776649.24080: variable 'ansible_shell_executable' from source: unknown 9786 1726776649.24082: variable 'ansible_host' from source: host vars for 'managed_node3' 9786 1726776649.24084: variable 'ansible_pipelining' from source: unknown 9786 1726776649.24086: variable 'ansible_timeout' from source: unknown 9786 1726776649.24088: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9786 1726776649.24163: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9786 1726776649.24174: variable 'omit' from source: magic vars 9786 1726776649.24178: starting attempt loop 9786 1726776649.24180: running the handler 9786 1726776649.24218: handler run complete 9786 1726776649.24225: attempt loop complete, returning result 9786 1726776649.24228: _execute() done 9786 1726776649.24232: dumping result to json 9786 1726776649.24235: done dumping result, returning 9786 1726776649.24239: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables [120fa90a-8a95-fee7-0cea-0000000000c3] 9786 1726776649.24243: sending task result for task 120fa90a-8a95-fee7-0cea-0000000000c3 9786 1726776649.24262: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000000c3 9786 1726776649.24264: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_packages": [ "tuned", "python3-configobj" ], "__kernel_settings_services": [ "tuned" ] }, "ansible_included_var_files": [ "/tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/vars/default.yml" ], "changed": false } 9704 1726776649.24504: no more pending results, returning what we have 9704 1726776649.24506: results queue empty 9704 1726776649.24506: checking for any_errors_fatal 9704 1726776649.24511: done checking for any_errors_fatal 9704 1726776649.24512: checking for max_fail_percentage 9704 1726776649.24512: done checking for max_fail_percentage 9704 1726776649.24513: checking to see if all hosts have failed and the running result is not ok 9704 1726776649.24513: done checking to see if all hosts have failed 9704 1726776649.24513: getting the remaining hosts for this loop 9704 1726776649.24514: done getting the remaining hosts for this loop 9704 1726776649.24516: getting the next task for host managed_node3 9704 1726776649.24521: done getting next task for host managed_node3 9704 1726776649.24523: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed 9704 1726776649.24525: ^ state is: HOST STATE: block=2, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776649.24533: getting variables 9704 1726776649.24534: in VariableManager get_vars() 9704 1726776649.24555: Calling all_inventory to load vars for managed_node3 9704 1726776649.24556: Calling groups_inventory to load vars for managed_node3 9704 1726776649.24557: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776649.24563: Calling all_plugins_play to load vars for managed_node3 9704 1726776649.24565: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776649.24567: Calling groups_plugins_play to load vars for managed_node3 9704 1726776649.24666: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776649.24784: done with get_vars() 9704 1726776649.24791: done getting variables 9704 1726776649.24860: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure required packages are installed] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:12 Thursday 19 September 2024 16:10:49 -0400 (0:00:00.027) 0:00:02.177 **** 9704 1726776649.24880: entering _queue_task() for managed_node3/package 9704 1726776649.24881: Creating lock for package 9704 1726776649.25032: worker is 1 (out of 1 available) 9704 1726776649.25044: exiting _queue_task() for managed_node3/package 9704 1726776649.25055: done queuing things up, now waiting for results queue to drain 9704 1726776649.25057: waiting for pending results... 9787 1726776649.25151: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed 9787 1726776649.25243: in run() - task 120fa90a-8a95-fee7-0cea-000000000049 9787 1726776649.25255: variable 'ansible_search_path' from source: unknown 9787 1726776649.25258: variable 'ansible_search_path' from source: unknown 9787 1726776649.25282: calling self._execute() 9787 1726776649.25334: variable 'ansible_host' from source: host vars for 'managed_node3' 9787 1726776649.25340: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9787 1726776649.25345: variable 'omit' from source: magic vars 9787 1726776649.25410: variable 'omit' from source: magic vars 9787 1726776649.25440: variable 'omit' from source: magic vars 9787 1726776649.25457: variable '__kernel_settings_packages' from source: include_vars 9787 1726776649.25662: variable '__kernel_settings_packages' from source: include_vars 9787 1726776649.25867: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9787 1726776649.27322: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9787 1726776649.27381: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9787 1726776649.27412: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9787 1726776649.27440: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9787 1726776649.27461: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9787 1726776649.27538: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9787 1726776649.27560: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9787 1726776649.27578: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9787 1726776649.27601: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9787 1726776649.27610: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9787 1726776649.27685: variable '__kernel_settings_is_ostree' from source: set_fact 9787 1726776649.27692: variable 'omit' from source: magic vars 9787 1726776649.27714: variable 'omit' from source: magic vars 9787 1726776649.27735: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9787 1726776649.27755: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9787 1726776649.27769: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9787 1726776649.27781: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9787 1726776649.27789: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9787 1726776649.27810: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9787 1726776649.27814: variable 'ansible_host' from source: host vars for 'managed_node3' 9787 1726776649.27816: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9787 1726776649.27894: Set connection var ansible_module_compression to ZIP_DEFLATED 9787 1726776649.27902: Set connection var ansible_shell_type to sh 9787 1726776649.27911: Set connection var ansible_timeout to 10 9787 1726776649.27917: Set connection var ansible_pipelining to False 9787 1726776649.27924: Set connection var ansible_shell_executable to /bin/sh 9787 1726776649.27927: Set connection var ansible_connection to ssh 9787 1726776649.27945: variable 'ansible_shell_executable' from source: unknown 9787 1726776649.27949: variable 'ansible_connection' from source: unknown 9787 1726776649.27953: variable 'ansible_module_compression' from source: unknown 9787 1726776649.27956: variable 'ansible_shell_type' from source: unknown 9787 1726776649.27960: variable 'ansible_shell_executable' from source: unknown 9787 1726776649.27963: variable 'ansible_host' from source: host vars for 'managed_node3' 9787 1726776649.27967: variable 'ansible_pipelining' from source: unknown 9787 1726776649.27971: variable 'ansible_timeout' from source: unknown 9787 1726776649.27974: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9787 1726776649.28039: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9787 1726776649.28050: variable 'omit' from source: magic vars 9787 1726776649.28056: starting attempt loop 9787 1726776649.28060: running the handler 9787 1726776649.28124: variable 'ansible_facts' from source: unknown 9787 1726776649.28206: _low_level_execute_command(): starting 9787 1726776649.28215: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9787 1726776649.30489: stdout chunk (state=2): >>>/root <<< 9787 1726776649.30615: stderr chunk (state=3): >>><<< 9787 1726776649.30623: stdout chunk (state=3): >>><<< 9787 1726776649.30644: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9787 1726776649.30659: _low_level_execute_command(): starting 9787 1726776649.30667: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338 `" && echo ansible-tmp-1726776649.306539-9787-195426086387338="` echo /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338 `" ) && sleep 0' 9787 1726776649.33131: stdout chunk (state=2): >>>ansible-tmp-1726776649.306539-9787-195426086387338=/root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338 <<< 9787 1726776649.33267: stderr chunk (state=3): >>><<< 9787 1726776649.33274: stdout chunk (state=3): >>><<< 9787 1726776649.33290: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776649.306539-9787-195426086387338=/root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338 , stderr= 9787 1726776649.33320: variable 'ansible_module_compression' from source: unknown 9787 1726776649.33370: ANSIBALLZ: Using generic lock for ansible.legacy.dnf 9787 1726776649.33375: ANSIBALLZ: Acquiring lock 9787 1726776649.33378: ANSIBALLZ: Lock acquired: 140007686719920 9787 1726776649.33382: ANSIBALLZ: Creating module 9787 1726776649.46030: ANSIBALLZ: Writing module into payload 9787 1726776649.46234: ANSIBALLZ: Writing module 9787 1726776649.46257: ANSIBALLZ: Renaming module 9787 1726776649.46264: ANSIBALLZ: Done creating module 9787 1726776649.46279: variable 'ansible_facts' from source: unknown 9787 1726776649.46362: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/AnsiballZ_dnf.py 9787 1726776649.46467: Sending initial data 9787 1726776649.46474: Sent initial data (149 bytes) 9787 1726776649.49025: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp6u3793na /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/AnsiballZ_dnf.py <<< 9787 1726776649.50298: stderr chunk (state=3): >>><<< 9787 1726776649.50307: stdout chunk (state=3): >>><<< 9787 1726776649.50328: done transferring module to remote 9787 1726776649.50341: _low_level_execute_command(): starting 9787 1726776649.50347: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/ /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/AnsiballZ_dnf.py && sleep 0' 9787 1726776649.52752: stderr chunk (state=2): >>><<< 9787 1726776649.52761: stdout chunk (state=2): >>><<< 9787 1726776649.52776: _low_level_execute_command() done: rc=0, stdout=, stderr= 9787 1726776649.52780: _low_level_execute_command(): starting 9787 1726776649.52785: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/AnsiballZ_dnf.py && sleep 0' 9787 1726776652.05438: stdout chunk (state=2): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["tuned", "python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} <<< 9787 1726776652.13219: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9787 1726776652.13269: stderr chunk (state=3): >>><<< 9787 1726776652.13278: stdout chunk (state=3): >>><<< 9787 1726776652.13295: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["tuned", "python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} , stderr=Shared connection to 10.31.8.186 closed. 9787 1726776652.13334: done with _execute_module (ansible.legacy.dnf, {'name': ['tuned', 'python3-configobj'], 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9787 1726776652.13343: _low_level_execute_command(): starting 9787 1726776652.13349: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776649.306539-9787-195426086387338/ > /dev/null 2>&1 && sleep 0' 9787 1726776652.15719: stderr chunk (state=2): >>><<< 9787 1726776652.15726: stdout chunk (state=2): >>><<< 9787 1726776652.15741: _low_level_execute_command() done: rc=0, stdout=, stderr= 9787 1726776652.15749: handler run complete 9787 1726776652.15775: attempt loop complete, returning result 9787 1726776652.15778: _execute() done 9787 1726776652.15780: dumping result to json 9787 1726776652.15784: done dumping result, returning 9787 1726776652.15789: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed [120fa90a-8a95-fee7-0cea-000000000049] 9787 1726776652.15792: sending task result for task 120fa90a-8a95-fee7-0cea-000000000049 9787 1726776652.15819: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000049 9787 1726776652.15822: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 9704 1726776652.16104: no more pending results, returning what we have 9704 1726776652.16107: results queue empty 9704 1726776652.16107: checking for any_errors_fatal 9704 1726776652.16113: done checking for any_errors_fatal 9704 1726776652.16113: checking for max_fail_percentage 9704 1726776652.16114: done checking for max_fail_percentage 9704 1726776652.16115: checking to see if all hosts have failed and the running result is not ok 9704 1726776652.16116: done checking to see if all hosts have failed 9704 1726776652.16116: getting the remaining hosts for this loop 9704 1726776652.16117: done getting the remaining hosts for this loop 9704 1726776652.16120: getting the next task for host managed_node3 9704 1726776652.16128: done getting next task for host managed_node3 9704 1726776652.16132: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes 9704 1726776652.16134: ^ state is: HOST STATE: block=2, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776652.16143: getting variables 9704 1726776652.16144: in VariableManager get_vars() 9704 1726776652.16167: Calling all_inventory to load vars for managed_node3 9704 1726776652.16170: Calling groups_inventory to load vars for managed_node3 9704 1726776652.16171: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776652.16177: Calling all_plugins_play to load vars for managed_node3 9704 1726776652.16179: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776652.16180: Calling groups_plugins_play to load vars for managed_node3 9704 1726776652.16317: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776652.16435: done with get_vars() 9704 1726776652.16443: done getting variables 9704 1726776652.16509: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:24 Thursday 19 September 2024 16:10:52 -0400 (0:00:02.916) 0:00:05.094 **** 9704 1726776652.16535: entering _queue_task() for managed_node3/debug 9704 1726776652.16536: Creating lock for debug 9704 1726776652.16698: worker is 1 (out of 1 available) 9704 1726776652.16712: exiting _queue_task() for managed_node3/debug 9704 1726776652.16724: done queuing things up, now waiting for results queue to drain 9704 1726776652.16726: waiting for pending results... 9841 1726776652.16829: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes 9841 1726776652.16937: in run() - task 120fa90a-8a95-fee7-0cea-00000000004b 9841 1726776652.16953: variable 'ansible_search_path' from source: unknown 9841 1726776652.16957: variable 'ansible_search_path' from source: unknown 9841 1726776652.16985: calling self._execute() 9841 1726776652.17044: variable 'ansible_host' from source: host vars for 'managed_node3' 9841 1726776652.17052: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9841 1726776652.17058: variable 'omit' from source: magic vars 9841 1726776652.17386: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9841 1726776652.18881: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9841 1726776652.18931: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9841 1726776652.18959: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9841 1726776652.19000: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9841 1726776652.19022: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9841 1726776652.19076: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9841 1726776652.19100: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9841 1726776652.19121: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9841 1726776652.19150: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9841 1726776652.19161: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9841 1726776652.19239: variable '__kernel_settings_is_transactional' from source: set_fact 9841 1726776652.19254: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 9841 1726776652.19259: when evaluation is False, skipping this task 9841 1726776652.19262: _execute() done 9841 1726776652.19266: dumping result to json 9841 1726776652.19269: done dumping result, returning 9841 1726776652.19275: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes [120fa90a-8a95-fee7-0cea-00000000004b] 9841 1726776652.19280: sending task result for task 120fa90a-8a95-fee7-0cea-00000000004b 9841 1726776652.19301: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000004b 9841 1726776652.19305: WORKER PROCESS EXITING skipping: [managed_node3] => { "false_condition": "__kernel_settings_is_transactional | d(false)" } 9704 1726776652.19405: no more pending results, returning what we have 9704 1726776652.19408: results queue empty 9704 1726776652.19408: checking for any_errors_fatal 9704 1726776652.19417: done checking for any_errors_fatal 9704 1726776652.19417: checking for max_fail_percentage 9704 1726776652.19419: done checking for max_fail_percentage 9704 1726776652.19419: checking to see if all hosts have failed and the running result is not ok 9704 1726776652.19420: done checking to see if all hosts have failed 9704 1726776652.19420: getting the remaining hosts for this loop 9704 1726776652.19421: done getting the remaining hosts for this loop 9704 1726776652.19424: getting the next task for host managed_node3 9704 1726776652.19430: done getting next task for host managed_node3 9704 1726776652.19433: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems 9704 1726776652.19435: ^ state is: HOST STATE: block=2, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776652.19447: getting variables 9704 1726776652.19448: in VariableManager get_vars() 9704 1726776652.19475: Calling all_inventory to load vars for managed_node3 9704 1726776652.19479: Calling groups_inventory to load vars for managed_node3 9704 1726776652.19481: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776652.19488: Calling all_plugins_play to load vars for managed_node3 9704 1726776652.19490: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776652.19492: Calling groups_plugins_play to load vars for managed_node3 9704 1726776652.19596: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776652.19709: done with get_vars() 9704 1726776652.19717: done getting variables 9704 1726776652.19812: Loading ActionModule 'reboot' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/reboot.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Reboot transactional update systems] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:29 Thursday 19 September 2024 16:10:52 -0400 (0:00:00.032) 0:00:05.127 **** 9704 1726776652.19836: entering _queue_task() for managed_node3/reboot 9704 1726776652.19837: Creating lock for reboot 9704 1726776652.19982: worker is 1 (out of 1 available) 9704 1726776652.19993: exiting _queue_task() for managed_node3/reboot 9704 1726776652.20002: done queuing things up, now waiting for results queue to drain 9704 1726776652.20003: waiting for pending results... 9842 1726776652.20160: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems 9842 1726776652.20266: in run() - task 120fa90a-8a95-fee7-0cea-00000000004c 9842 1726776652.20282: variable 'ansible_search_path' from source: unknown 9842 1726776652.20286: variable 'ansible_search_path' from source: unknown 9842 1726776652.20315: calling self._execute() 9842 1726776652.20369: variable 'ansible_host' from source: host vars for 'managed_node3' 9842 1726776652.20377: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9842 1726776652.20385: variable 'omit' from source: magic vars 9842 1726776652.20770: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9842 1726776652.22241: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9842 1726776652.22295: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9842 1726776652.22326: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9842 1726776652.22353: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9842 1726776652.22371: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9842 1726776652.22426: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9842 1726776652.22458: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9842 1726776652.22478: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9842 1726776652.22506: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9842 1726776652.22519: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9842 1726776652.22599: variable '__kernel_settings_is_transactional' from source: set_fact 9842 1726776652.22617: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 9842 1726776652.22620: when evaluation is False, skipping this task 9842 1726776652.22623: _execute() done 9842 1726776652.22625: dumping result to json 9842 1726776652.22627: done dumping result, returning 9842 1726776652.22632: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems [120fa90a-8a95-fee7-0cea-00000000004c] 9842 1726776652.22637: sending task result for task 120fa90a-8a95-fee7-0cea-00000000004c 9842 1726776652.22657: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000004c 9842 1726776652.22659: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_is_transactional | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776652.22955: no more pending results, returning what we have 9704 1726776652.22958: results queue empty 9704 1726776652.22958: checking for any_errors_fatal 9704 1726776652.22962: done checking for any_errors_fatal 9704 1726776652.22963: checking for max_fail_percentage 9704 1726776652.22964: done checking for max_fail_percentage 9704 1726776652.22965: checking to see if all hosts have failed and the running result is not ok 9704 1726776652.22965: done checking to see if all hosts have failed 9704 1726776652.22966: getting the remaining hosts for this loop 9704 1726776652.22967: done getting the remaining hosts for this loop 9704 1726776652.22970: getting the next task for host managed_node3 9704 1726776652.22974: done getting next task for host managed_node3 9704 1726776652.22976: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set 9704 1726776652.22978: ^ state is: HOST STATE: block=2, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776652.22989: getting variables 9704 1726776652.22990: in VariableManager get_vars() 9704 1726776652.23063: Calling all_inventory to load vars for managed_node3 9704 1726776652.23065: Calling groups_inventory to load vars for managed_node3 9704 1726776652.23066: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776652.23073: Calling all_plugins_play to load vars for managed_node3 9704 1726776652.23074: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776652.23076: Calling groups_plugins_play to load vars for managed_node3 9704 1726776652.23173: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776652.23286: done with get_vars() 9704 1726776652.23294: done getting variables 9704 1726776652.23338: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:34 Thursday 19 September 2024 16:10:52 -0400 (0:00:00.035) 0:00:05.162 **** 9704 1726776652.23358: entering _queue_task() for managed_node3/fail 9704 1726776652.23530: worker is 1 (out of 1 available) 9704 1726776652.23544: exiting _queue_task() for managed_node3/fail 9704 1726776652.23556: done queuing things up, now waiting for results queue to drain 9704 1726776652.23557: waiting for pending results... 9843 1726776652.23663: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set 9843 1726776652.23777: in run() - task 120fa90a-8a95-fee7-0cea-00000000004d 9843 1726776652.23792: variable 'ansible_search_path' from source: unknown 9843 1726776652.23797: variable 'ansible_search_path' from source: unknown 9843 1726776652.23826: calling self._execute() 9843 1726776652.23882: variable 'ansible_host' from source: host vars for 'managed_node3' 9843 1726776652.23889: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9843 1726776652.23895: variable 'omit' from source: magic vars 9843 1726776652.24241: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9843 1726776652.25744: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9843 1726776652.25791: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9843 1726776652.25821: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9843 1726776652.25857: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9843 1726776652.25878: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9843 1726776652.25937: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9843 1726776652.25957: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9843 1726776652.25973: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9843 1726776652.25995: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9843 1726776652.26005: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9843 1726776652.26090: variable '__kernel_settings_is_transactional' from source: set_fact 9843 1726776652.26107: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 9843 1726776652.26111: when evaluation is False, skipping this task 9843 1726776652.26118: _execute() done 9843 1726776652.26121: dumping result to json 9843 1726776652.26123: done dumping result, returning 9843 1726776652.26127: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set [120fa90a-8a95-fee7-0cea-00000000004d] 9843 1726776652.26132: sending task result for task 120fa90a-8a95-fee7-0cea-00000000004d 9843 1726776652.26151: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000004d 9843 1726776652.26153: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_is_transactional | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776652.26346: no more pending results, returning what we have 9704 1726776652.26349: results queue empty 9704 1726776652.26350: checking for any_errors_fatal 9704 1726776652.26353: done checking for any_errors_fatal 9704 1726776652.26354: checking for max_fail_percentage 9704 1726776652.26355: done checking for max_fail_percentage 9704 1726776652.26355: checking to see if all hosts have failed and the running result is not ok 9704 1726776652.26356: done checking to see if all hosts have failed 9704 1726776652.26356: getting the remaining hosts for this loop 9704 1726776652.26357: done getting the remaining hosts for this loop 9704 1726776652.26359: getting the next task for host managed_node3 9704 1726776652.26365: done getting next task for host managed_node3 9704 1726776652.26367: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config 9704 1726776652.26368: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776652.26377: getting variables 9704 1726776652.26378: in VariableManager get_vars() 9704 1726776652.26399: Calling all_inventory to load vars for managed_node3 9704 1726776652.26401: Calling groups_inventory to load vars for managed_node3 9704 1726776652.26402: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776652.26407: Calling all_plugins_play to load vars for managed_node3 9704 1726776652.26409: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776652.26410: Calling groups_plugins_play to load vars for managed_node3 9704 1726776652.26510: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776652.26625: done with get_vars() 9704 1726776652.26635: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Read tuned main config] ****** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:42 Thursday 19 September 2024 16:10:52 -0400 (0:00:00.033) 0:00:05.195 **** 9704 1726776652.26690: entering _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776652.26691: Creating lock for fedora.linux_system_roles.kernel_settings_get_config 9704 1726776652.26854: worker is 1 (out of 1 available) 9704 1726776652.26867: exiting _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776652.26879: done queuing things up, now waiting for results queue to drain 9704 1726776652.26881: waiting for pending results... 9844 1726776652.26986: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config 9844 1726776652.27089: in run() - task 120fa90a-8a95-fee7-0cea-00000000004f 9844 1726776652.27104: variable 'ansible_search_path' from source: unknown 9844 1726776652.27108: variable 'ansible_search_path' from source: unknown 9844 1726776652.27139: calling self._execute() 9844 1726776652.27246: variable 'ansible_host' from source: host vars for 'managed_node3' 9844 1726776652.27254: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9844 1726776652.27263: variable 'omit' from source: magic vars 9844 1726776652.27335: variable 'omit' from source: magic vars 9844 1726776652.27368: variable 'omit' from source: magic vars 9844 1726776652.27389: variable '__kernel_settings_tuned_main_conf_file' from source: role '' all vars 9844 1726776652.27587: variable '__kernel_settings_tuned_main_conf_file' from source: role '' all vars 9844 1726776652.27645: variable '__kernel_settings_tuned_dir' from source: role '' all vars 9844 1726776652.27672: variable 'omit' from source: magic vars 9844 1726776652.27700: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9844 1726776652.27725: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9844 1726776652.27742: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9844 1726776652.27753: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9844 1726776652.27761: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9844 1726776652.27783: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9844 1726776652.27787: variable 'ansible_host' from source: host vars for 'managed_node3' 9844 1726776652.27790: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9844 1726776652.27866: Set connection var ansible_module_compression to ZIP_DEFLATED 9844 1726776652.27874: Set connection var ansible_shell_type to sh 9844 1726776652.27882: Set connection var ansible_timeout to 10 9844 1726776652.27888: Set connection var ansible_pipelining to False 9844 1726776652.27896: Set connection var ansible_shell_executable to /bin/sh 9844 1726776652.27899: Set connection var ansible_connection to ssh 9844 1726776652.27917: variable 'ansible_shell_executable' from source: unknown 9844 1726776652.27921: variable 'ansible_connection' from source: unknown 9844 1726776652.27924: variable 'ansible_module_compression' from source: unknown 9844 1726776652.27929: variable 'ansible_shell_type' from source: unknown 9844 1726776652.27933: variable 'ansible_shell_executable' from source: unknown 9844 1726776652.27936: variable 'ansible_host' from source: host vars for 'managed_node3' 9844 1726776652.27941: variable 'ansible_pipelining' from source: unknown 9844 1726776652.27944: variable 'ansible_timeout' from source: unknown 9844 1726776652.27948: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9844 1726776652.28063: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 9844 1726776652.28074: variable 'omit' from source: magic vars 9844 1726776652.28079: starting attempt loop 9844 1726776652.28083: running the handler 9844 1726776652.28092: _low_level_execute_command(): starting 9844 1726776652.28098: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9844 1726776652.30397: stdout chunk (state=2): >>>/root <<< 9844 1726776652.30516: stderr chunk (state=3): >>><<< 9844 1726776652.30523: stdout chunk (state=3): >>><<< 9844 1726776652.30543: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9844 1726776652.30558: _low_level_execute_command(): starting 9844 1726776652.30564: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197 `" && echo ansible-tmp-1726776652.305532-9844-207190182256197="` echo /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197 `" ) && sleep 0' 9844 1726776652.33001: stdout chunk (state=2): >>>ansible-tmp-1726776652.305532-9844-207190182256197=/root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197 <<< 9844 1726776652.33133: stderr chunk (state=3): >>><<< 9844 1726776652.33140: stdout chunk (state=3): >>><<< 9844 1726776652.33155: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776652.305532-9844-207190182256197=/root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197 , stderr= 9844 1726776652.33191: variable 'ansible_module_compression' from source: unknown 9844 1726776652.33224: ANSIBALLZ: Using lock for fedora.linux_system_roles.kernel_settings_get_config 9844 1726776652.33229: ANSIBALLZ: Acquiring lock 9844 1726776652.33234: ANSIBALLZ: Lock acquired: 140007684611808 9844 1726776652.33238: ANSIBALLZ: Creating module 9844 1726776652.44103: ANSIBALLZ: Writing module into payload 9844 1726776652.44165: ANSIBALLZ: Writing module 9844 1726776652.44187: ANSIBALLZ: Renaming module 9844 1726776652.44194: ANSIBALLZ: Done creating module 9844 1726776652.44211: variable 'ansible_facts' from source: unknown 9844 1726776652.44272: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/AnsiballZ_kernel_settings_get_config.py 9844 1726776652.44372: Sending initial data 9844 1726776652.44379: Sent initial data (172 bytes) 9844 1726776652.46883: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpfkpofc05 /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/AnsiballZ_kernel_settings_get_config.py <<< 9844 1726776652.47875: stderr chunk (state=3): >>><<< 9844 1726776652.47882: stdout chunk (state=3): >>><<< 9844 1726776652.47902: done transferring module to remote 9844 1726776652.47912: _low_level_execute_command(): starting 9844 1726776652.47920: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/ /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/AnsiballZ_kernel_settings_get_config.py && sleep 0' 9844 1726776652.50222: stderr chunk (state=2): >>><<< 9844 1726776652.50232: stdout chunk (state=2): >>><<< 9844 1726776652.50246: _low_level_execute_command() done: rc=0, stdout=, stderr= 9844 1726776652.50251: _low_level_execute_command(): starting 9844 1726776652.50256: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/AnsiballZ_kernel_settings_get_config.py && sleep 0' 9844 1726776652.66183: stdout chunk (state=2): >>> {"changed": false, "data": {"daemon": "1", "dynamic_tuning": "0", "sleep_interval": "1", "update_interval": "10", "recommend_command": "1", "reapply_sysctl": "1", "default_instance_priority": "0", "udev_buffer_size": "1MB", "log_file_count": "2", "log_file_max_size": "1MB"}, "invocation": {"module_args": {"path": "/etc/tuned/tuned-main.conf"}}} <<< 9844 1726776652.67277: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9844 1726776652.67331: stderr chunk (state=3): >>><<< 9844 1726776652.67338: stdout chunk (state=3): >>><<< 9844 1726776652.67356: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "data": {"daemon": "1", "dynamic_tuning": "0", "sleep_interval": "1", "update_interval": "10", "recommend_command": "1", "reapply_sysctl": "1", "default_instance_priority": "0", "udev_buffer_size": "1MB", "log_file_count": "2", "log_file_max_size": "1MB"}, "invocation": {"module_args": {"path": "/etc/tuned/tuned-main.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 9844 1726776652.67382: done with _execute_module (fedora.linux_system_roles.kernel_settings_get_config, {'path': '/etc/tuned/tuned-main.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'fedora.linux_system_roles.kernel_settings_get_config', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9844 1726776652.67393: _low_level_execute_command(): starting 9844 1726776652.67398: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776652.305532-9844-207190182256197/ > /dev/null 2>&1 && sleep 0' 9844 1726776652.69799: stderr chunk (state=2): >>><<< 9844 1726776652.69807: stdout chunk (state=2): >>><<< 9844 1726776652.69824: _low_level_execute_command() done: rc=0, stdout=, stderr= 9844 1726776652.69833: handler run complete 9844 1726776652.69846: attempt loop complete, returning result 9844 1726776652.69850: _execute() done 9844 1726776652.69854: dumping result to json 9844 1726776652.69858: done dumping result, returning 9844 1726776652.69865: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config [120fa90a-8a95-fee7-0cea-00000000004f] 9844 1726776652.69870: sending task result for task 120fa90a-8a95-fee7-0cea-00000000004f 9844 1726776652.69901: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000004f 9844 1726776652.69904: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "data": { "daemon": "1", "default_instance_priority": "0", "dynamic_tuning": "0", "log_file_count": "2", "log_file_max_size": "1MB", "reapply_sysctl": "1", "recommend_command": "1", "sleep_interval": "1", "udev_buffer_size": "1MB", "update_interval": "10" } } 9704 1726776652.70231: no more pending results, returning what we have 9704 1726776652.70234: results queue empty 9704 1726776652.70234: checking for any_errors_fatal 9704 1726776652.70238: done checking for any_errors_fatal 9704 1726776652.70238: checking for max_fail_percentage 9704 1726776652.70239: done checking for max_fail_percentage 9704 1726776652.70239: checking to see if all hosts have failed and the running result is not ok 9704 1726776652.70240: done checking to see if all hosts have failed 9704 1726776652.70240: getting the remaining hosts for this loop 9704 1726776652.70241: done getting the remaining hosts for this loop 9704 1726776652.70243: getting the next task for host managed_node3 9704 1726776652.70247: done getting next task for host managed_node3 9704 1726776652.70249: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory 9704 1726776652.70251: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776652.70256: getting variables 9704 1726776652.70257: in VariableManager get_vars() 9704 1726776652.70277: Calling all_inventory to load vars for managed_node3 9704 1726776652.70278: Calling groups_inventory to load vars for managed_node3 9704 1726776652.70279: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776652.70285: Calling all_plugins_play to load vars for managed_node3 9704 1726776652.70287: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776652.70288: Calling groups_plugins_play to load vars for managed_node3 9704 1726776652.70386: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776652.70500: done with get_vars() 9704 1726776652.70508: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:50 Thursday 19 September 2024 16:10:52 -0400 (0:00:00.438) 0:00:05.634 **** 9704 1726776652.70575: entering _queue_task() for managed_node3/stat 9704 1726776652.70751: worker is 1 (out of 1 available) 9704 1726776652.70766: exiting _queue_task() for managed_node3/stat 9704 1726776652.70777: done queuing things up, now waiting for results queue to drain 9704 1726776652.70780: waiting for pending results... 9863 1726776652.70885: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory 9863 1726776652.70993: in run() - task 120fa90a-8a95-fee7-0cea-000000000050 9863 1726776652.71009: variable 'ansible_search_path' from source: unknown 9863 1726776652.71016: variable 'ansible_search_path' from source: unknown 9863 1726776652.71054: variable '__prof_from_conf' from source: task vars 9863 1726776652.71273: variable '__prof_from_conf' from source: task vars 9863 1726776652.71402: variable '__data' from source: task vars 9863 1726776652.71461: variable '__kernel_settings_register_tuned_main' from source: set_fact 9863 1726776652.71597: variable '__kernel_settings_tuned_dir' from source: role '' all vars 9863 1726776652.71607: variable '__kernel_settings_tuned_dir' from source: role '' all vars 9863 1726776652.71652: variable '__kernel_settings_tuned_dir' from source: role '' all vars 9863 1726776652.71669: variable 'omit' from source: magic vars 9863 1726776652.71771: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776652.71782: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776652.71792: variable 'omit' from source: magic vars 9863 1726776652.72069: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9863 1726776652.73673: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9863 1726776652.73727: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9863 1726776652.73756: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9863 1726776652.73783: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9863 1726776652.73805: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9863 1726776652.73861: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9863 1726776652.73886: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9863 1726776652.73910: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9863 1726776652.73942: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9863 1726776652.73954: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9863 1726776652.74022: variable 'item' from source: unknown 9863 1726776652.74036: Evaluated conditional (item | length > 0): False 9863 1726776652.74040: when evaluation is False, skipping this task 9863 1726776652.74061: variable 'item' from source: unknown 9863 1726776652.74106: variable 'item' from source: unknown skipping: [managed_node3] => (item=) => { "ansible_loop_var": "item", "changed": false, "false_condition": "item | length > 0", "item": "", "skip_reason": "Conditional result was False" } 9863 1726776652.74182: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776652.74193: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776652.74202: variable 'omit' from source: magic vars 9863 1726776652.74318: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9863 1726776652.74338: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9863 1726776652.74356: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9863 1726776652.74384: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9863 1726776652.74396: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9863 1726776652.74469: variable 'item' from source: unknown 9863 1726776652.74478: Evaluated conditional (item | length > 0): True 9863 1726776652.74484: variable 'omit' from source: magic vars 9863 1726776652.74521: variable 'omit' from source: magic vars 9863 1726776652.74566: variable 'item' from source: unknown 9863 1726776652.74627: variable 'item' from source: unknown 9863 1726776652.74645: variable 'omit' from source: magic vars 9863 1726776652.74667: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9863 1726776652.74688: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9863 1726776652.74702: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9863 1726776652.74717: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9863 1726776652.74726: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9863 1726776652.74751: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9863 1726776652.74755: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776652.74758: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776652.74843: Set connection var ansible_module_compression to ZIP_DEFLATED 9863 1726776652.74850: Set connection var ansible_shell_type to sh 9863 1726776652.74857: Set connection var ansible_timeout to 10 9863 1726776652.74862: Set connection var ansible_pipelining to False 9863 1726776652.74868: Set connection var ansible_shell_executable to /bin/sh 9863 1726776652.74870: Set connection var ansible_connection to ssh 9863 1726776652.74885: variable 'ansible_shell_executable' from source: unknown 9863 1726776652.74889: variable 'ansible_connection' from source: unknown 9863 1726776652.74892: variable 'ansible_module_compression' from source: unknown 9863 1726776652.74895: variable 'ansible_shell_type' from source: unknown 9863 1726776652.74898: variable 'ansible_shell_executable' from source: unknown 9863 1726776652.74901: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776652.74905: variable 'ansible_pipelining' from source: unknown 9863 1726776652.74908: variable 'ansible_timeout' from source: unknown 9863 1726776652.74912: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776652.75083: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 9863 1726776652.75095: variable 'omit' from source: magic vars 9863 1726776652.75100: starting attempt loop 9863 1726776652.75104: running the handler 9863 1726776652.75118: _low_level_execute_command(): starting 9863 1726776652.75127: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9863 1726776652.77786: stdout chunk (state=2): >>>/root <<< 9863 1726776652.77906: stderr chunk (state=3): >>><<< 9863 1726776652.77913: stdout chunk (state=3): >>><<< 9863 1726776652.77932: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9863 1726776652.77944: _low_level_execute_command(): starting 9863 1726776652.77948: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173 `" && echo ansible-tmp-1726776652.7793958-9863-109415551609173="` echo /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173 `" ) && sleep 0' 9863 1726776652.80368: stdout chunk (state=2): >>>ansible-tmp-1726776652.7793958-9863-109415551609173=/root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173 <<< 9863 1726776652.80490: stderr chunk (state=3): >>><<< 9863 1726776652.80498: stdout chunk (state=3): >>><<< 9863 1726776652.80511: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776652.7793958-9863-109415551609173=/root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173 , stderr= 9863 1726776652.80551: variable 'ansible_module_compression' from source: unknown 9863 1726776652.80591: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 9863 1726776652.80621: variable 'ansible_facts' from source: unknown 9863 1726776652.80687: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/AnsiballZ_stat.py 9863 1726776652.80785: Sending initial data 9863 1726776652.80792: Sent initial data (151 bytes) 9863 1726776652.83222: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpiqj0h9m1 /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/AnsiballZ_stat.py <<< 9863 1726776652.84193: stderr chunk (state=3): >>><<< 9863 1726776652.84200: stdout chunk (state=3): >>><<< 9863 1726776652.84221: done transferring module to remote 9863 1726776652.84233: _low_level_execute_command(): starting 9863 1726776652.84239: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/ /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/AnsiballZ_stat.py && sleep 0' 9863 1726776652.86537: stderr chunk (state=2): >>><<< 9863 1726776652.86544: stdout chunk (state=2): >>><<< 9863 1726776652.86558: _low_level_execute_command() done: rc=0, stdout=, stderr= 9863 1726776652.86563: _low_level_execute_command(): starting 9863 1726776652.86568: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/AnsiballZ_stat.py && sleep 0' 9863 1726776653.01654: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/profiles", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 9863 1726776653.02794: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9863 1726776653.02804: stdout chunk (state=3): >>><<< 9863 1726776653.02815: stderr chunk (state=3): >>><<< 9863 1726776653.02828: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/profiles", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 9863 1726776653.02852: done with _execute_module (stat, {'path': '/etc/tuned/profiles', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9863 1726776653.02864: _low_level_execute_command(): starting 9863 1726776653.02869: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776652.7793958-9863-109415551609173/ > /dev/null 2>&1 && sleep 0' 9863 1726776653.05292: stderr chunk (state=2): >>><<< 9863 1726776653.05301: stdout chunk (state=2): >>><<< 9863 1726776653.05315: _low_level_execute_command() done: rc=0, stdout=, stderr= 9863 1726776653.05325: handler run complete 9863 1726776653.05343: attempt loop complete, returning result 9863 1726776653.05359: variable 'item' from source: unknown 9863 1726776653.05419: variable 'item' from source: unknown ok: [managed_node3] => (item=/etc/tuned/profiles) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/tuned/profiles", "stat": { "exists": false } } 9863 1726776653.05506: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776653.05516: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776653.05525: variable 'omit' from source: magic vars 9863 1726776653.05637: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9863 1726776653.05660: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9863 1726776653.05679: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9863 1726776653.05705: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9863 1726776653.05717: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9863 1726776653.05776: variable 'item' from source: unknown 9863 1726776653.05785: Evaluated conditional (item | length > 0): True 9863 1726776653.05790: variable 'omit' from source: magic vars 9863 1726776653.05801: variable 'omit' from source: magic vars 9863 1726776653.05830: variable 'item' from source: unknown 9863 1726776653.05875: variable 'item' from source: unknown 9863 1726776653.05888: variable 'omit' from source: magic vars 9863 1726776653.05904: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9863 1726776653.05912: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9863 1726776653.05919: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9863 1726776653.05932: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9863 1726776653.05937: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776653.05940: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776653.05989: Set connection var ansible_module_compression to ZIP_DEFLATED 9863 1726776653.05995: Set connection var ansible_shell_type to sh 9863 1726776653.06003: Set connection var ansible_timeout to 10 9863 1726776653.06008: Set connection var ansible_pipelining to False 9863 1726776653.06015: Set connection var ansible_shell_executable to /bin/sh 9863 1726776653.06018: Set connection var ansible_connection to ssh 9863 1726776653.06035: variable 'ansible_shell_executable' from source: unknown 9863 1726776653.06039: variable 'ansible_connection' from source: unknown 9863 1726776653.06042: variable 'ansible_module_compression' from source: unknown 9863 1726776653.06045: variable 'ansible_shell_type' from source: unknown 9863 1726776653.06049: variable 'ansible_shell_executable' from source: unknown 9863 1726776653.06052: variable 'ansible_host' from source: host vars for 'managed_node3' 9863 1726776653.06056: variable 'ansible_pipelining' from source: unknown 9863 1726776653.06059: variable 'ansible_timeout' from source: unknown 9863 1726776653.06063: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9863 1726776653.06134: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9863 1726776653.06143: variable 'omit' from source: magic vars 9863 1726776653.06148: starting attempt loop 9863 1726776653.06152: running the handler 9863 1726776653.06158: _low_level_execute_command(): starting 9863 1726776653.06162: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9863 1726776653.08308: stdout chunk (state=2): >>>/root <<< 9863 1726776653.08697: stderr chunk (state=3): >>><<< 9863 1726776653.08705: stdout chunk (state=3): >>><<< 9863 1726776653.08726: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9863 1726776653.08739: _low_level_execute_command(): starting 9863 1726776653.08745: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516 `" && echo ansible-tmp-1726776653.087351-9863-19196872109516="` echo /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516 `" ) && sleep 0' 9863 1726776653.11422: stdout chunk (state=2): >>>ansible-tmp-1726776653.087351-9863-19196872109516=/root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516 <<< 9863 1726776653.11796: stderr chunk (state=3): >>><<< 9863 1726776653.11804: stdout chunk (state=3): >>><<< 9863 1726776653.11819: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776653.087351-9863-19196872109516=/root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516 , stderr= 9863 1726776653.11855: variable 'ansible_module_compression' from source: unknown 9863 1726776653.11899: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 9863 1726776653.11921: variable 'ansible_facts' from source: unknown 9863 1726776653.12002: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/AnsiballZ_stat.py 9863 1726776653.12464: Sending initial data 9863 1726776653.12471: Sent initial data (149 bytes) 9863 1726776653.15136: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpxcg81eki /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/AnsiballZ_stat.py <<< 9863 1726776653.16174: stderr chunk (state=3): >>><<< 9863 1726776653.16182: stdout chunk (state=3): >>><<< 9863 1726776653.16202: done transferring module to remote 9863 1726776653.16211: _low_level_execute_command(): starting 9863 1726776653.16216: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/ /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/AnsiballZ_stat.py && sleep 0' 9863 1726776653.18586: stderr chunk (state=2): >>><<< 9863 1726776653.18592: stdout chunk (state=2): >>><<< 9863 1726776653.18606: _low_level_execute_command() done: rc=0, stdout=, stderr= 9863 1726776653.18610: _low_level_execute_command(): starting 9863 1726776653.18615: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/AnsiballZ_stat.py && sleep 0' 9863 1726776653.34343: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned", "mode": "0755", "isdir": true, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 136, "inode": 917919, "dev": 51713, "nlink": 3, "atime": 1726776631.537892, "mtime": 1726776643.6350489, "ctime": 1726776643.6350489, "wusr": true, "rusr": true, "xusr": true, "wgrp": false, "rgrp": true, "xgrp": true, "woth": false, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "pw_name": "root", "gr_name": "root", "mimetype": "inode/directory", "charset": "binary", "version": "1785990601", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 9863 1726776653.35500: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9863 1726776653.35549: stderr chunk (state=3): >>><<< 9863 1726776653.35556: stdout chunk (state=3): >>><<< 9863 1726776653.35573: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned", "mode": "0755", "isdir": true, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 136, "inode": 917919, "dev": 51713, "nlink": 3, "atime": 1726776631.537892, "mtime": 1726776643.6350489, "ctime": 1726776643.6350489, "wusr": true, "rusr": true, "xusr": true, "wgrp": false, "rgrp": true, "xgrp": true, "woth": false, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "pw_name": "root", "gr_name": "root", "mimetype": "inode/directory", "charset": "binary", "version": "1785990601", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 9863 1726776653.35631: done with _execute_module (stat, {'path': '/etc/tuned', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9863 1726776653.35642: _low_level_execute_command(): starting 9863 1726776653.35648: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776653.087351-9863-19196872109516/ > /dev/null 2>&1 && sleep 0' 9863 1726776653.38037: stderr chunk (state=2): >>><<< 9863 1726776653.38045: stdout chunk (state=2): >>><<< 9863 1726776653.38061: _low_level_execute_command() done: rc=0, stdout=, stderr= 9863 1726776653.38068: handler run complete 9863 1726776653.38098: attempt loop complete, returning result 9863 1726776653.38114: variable 'item' from source: unknown 9863 1726776653.38177: variable 'item' from source: unknown ok: [managed_node3] => (item=/etc/tuned) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/tuned", "stat": { "atime": 1726776631.537892, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1726776643.6350489, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 917919, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/directory", "mode": "0755", "mtime": 1726776643.6350489, "nlink": 3, "path": "/etc/tuned", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 136, "uid": 0, "version": "1785990601", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 9863 1726776653.38221: dumping result to json 9863 1726776653.38233: done dumping result, returning 9863 1726776653.38242: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory [120fa90a-8a95-fee7-0cea-000000000050] 9863 1726776653.38247: sending task result for task 120fa90a-8a95-fee7-0cea-000000000050 9863 1726776653.38286: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000050 9863 1726776653.38290: WORKER PROCESS EXITING 9704 1726776653.38455: no more pending results, returning what we have 9704 1726776653.38458: results queue empty 9704 1726776653.38459: checking for any_errors_fatal 9704 1726776653.38463: done checking for any_errors_fatal 9704 1726776653.38463: checking for max_fail_percentage 9704 1726776653.38464: done checking for max_fail_percentage 9704 1726776653.38465: checking to see if all hosts have failed and the running result is not ok 9704 1726776653.38465: done checking to see if all hosts have failed 9704 1726776653.38466: getting the remaining hosts for this loop 9704 1726776653.38467: done getting the remaining hosts for this loop 9704 1726776653.38470: getting the next task for host managed_node3 9704 1726776653.38474: done getting next task for host managed_node3 9704 1726776653.38477: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir 9704 1726776653.38479: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776653.38489: getting variables 9704 1726776653.38490: in VariableManager get_vars() 9704 1726776653.38519: Calling all_inventory to load vars for managed_node3 9704 1726776653.38522: Calling groups_inventory to load vars for managed_node3 9704 1726776653.38524: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776653.38533: Calling all_plugins_play to load vars for managed_node3 9704 1726776653.38536: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776653.38538: Calling groups_plugins_play to load vars for managed_node3 9704 1726776653.38673: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776653.38784: done with get_vars() 9704 1726776653.38792: done getting variables 9704 1726776653.38836: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:63 Thursday 19 September 2024 16:10:53 -0400 (0:00:00.682) 0:00:06.317 **** 9704 1726776653.38857: entering _queue_task() for managed_node3/set_fact 9704 1726776653.39023: worker is 1 (out of 1 available) 9704 1726776653.39037: exiting _queue_task() for managed_node3/set_fact 9704 1726776653.39049: done queuing things up, now waiting for results queue to drain 9704 1726776653.39050: waiting for pending results... 9895 1726776653.39160: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir 9895 1726776653.39257: in run() - task 120fa90a-8a95-fee7-0cea-000000000051 9895 1726776653.39273: variable 'ansible_search_path' from source: unknown 9895 1726776653.39277: variable 'ansible_search_path' from source: unknown 9895 1726776653.39303: calling self._execute() 9895 1726776653.39364: variable 'ansible_host' from source: host vars for 'managed_node3' 9895 1726776653.39373: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9895 1726776653.39383: variable 'omit' from source: magic vars 9895 1726776653.39459: variable 'omit' from source: magic vars 9895 1726776653.39494: variable 'omit' from source: magic vars 9895 1726776653.39808: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9895 1726776653.41368: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9895 1726776653.41418: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9895 1726776653.41451: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9895 1726776653.41488: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9895 1726776653.41510: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9895 1726776653.41571: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9895 1726776653.41594: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9895 1726776653.41613: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9895 1726776653.41640: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9895 1726776653.41649: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9895 1726776653.41682: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9895 1726776653.41696: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9895 1726776653.41710: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9895 1726776653.41736: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9895 1726776653.41745: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9895 1726776653.41785: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9895 1726776653.41801: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9895 1726776653.41815: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9895 1726776653.41840: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9895 1726776653.41848: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9895 1726776653.42012: variable '__kernel_settings_find_profile_dirs' from source: set_fact 9895 1726776653.42087: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 9895 1726776653.42198: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 9895 1726776653.42231: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 9895 1726776653.42254: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 9895 1726776653.42275: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 9895 1726776653.42323: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 9895 1726776653.42344: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 9895 1726776653.42366: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 9895 1726776653.42388: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 9895 1726776653.42437: variable 'omit' from source: magic vars 9895 1726776653.42464: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9895 1726776653.42488: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9895 1726776653.42506: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9895 1726776653.42522: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9895 1726776653.42533: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9895 1726776653.42562: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9895 1726776653.42567: variable 'ansible_host' from source: host vars for 'managed_node3' 9895 1726776653.42572: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9895 1726776653.42667: Set connection var ansible_module_compression to ZIP_DEFLATED 9895 1726776653.42676: Set connection var ansible_shell_type to sh 9895 1726776653.42685: Set connection var ansible_timeout to 10 9895 1726776653.42691: Set connection var ansible_pipelining to False 9895 1726776653.42698: Set connection var ansible_shell_executable to /bin/sh 9895 1726776653.42701: Set connection var ansible_connection to ssh 9895 1726776653.42721: variable 'ansible_shell_executable' from source: unknown 9895 1726776653.43020: variable 'ansible_connection' from source: unknown 9895 1726776653.43026: variable 'ansible_module_compression' from source: unknown 9895 1726776653.43032: variable 'ansible_shell_type' from source: unknown 9895 1726776653.43035: variable 'ansible_shell_executable' from source: unknown 9895 1726776653.43038: variable 'ansible_host' from source: host vars for 'managed_node3' 9895 1726776653.43042: variable 'ansible_pipelining' from source: unknown 9895 1726776653.43044: variable 'ansible_timeout' from source: unknown 9895 1726776653.43048: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9895 1726776653.43180: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9895 1726776653.43193: variable 'omit' from source: magic vars 9895 1726776653.43199: starting attempt loop 9895 1726776653.43202: running the handler 9895 1726776653.43213: handler run complete 9895 1726776653.43221: attempt loop complete, returning result 9895 1726776653.43225: _execute() done 9895 1726776653.43228: dumping result to json 9895 1726776653.43232: done dumping result, returning 9895 1726776653.43238: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir [120fa90a-8a95-fee7-0cea-000000000051] 9895 1726776653.43243: sending task result for task 120fa90a-8a95-fee7-0cea-000000000051 9895 1726776653.43268: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000051 9895 1726776653.43271: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_profile_parent": "/etc/tuned" }, "changed": false } 9704 1726776653.43471: no more pending results, returning what we have 9704 1726776653.43474: results queue empty 9704 1726776653.43474: checking for any_errors_fatal 9704 1726776653.43481: done checking for any_errors_fatal 9704 1726776653.43482: checking for max_fail_percentage 9704 1726776653.43483: done checking for max_fail_percentage 9704 1726776653.43484: checking to see if all hosts have failed and the running result is not ok 9704 1726776653.43485: done checking to see if all hosts have failed 9704 1726776653.43485: getting the remaining hosts for this loop 9704 1726776653.43487: done getting the remaining hosts for this loop 9704 1726776653.43489: getting the next task for host managed_node3 9704 1726776653.43494: done getting next task for host managed_node3 9704 1726776653.43497: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started 9704 1726776653.43499: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776653.43508: getting variables 9704 1726776653.43509: in VariableManager get_vars() 9704 1726776653.43545: Calling all_inventory to load vars for managed_node3 9704 1726776653.43548: Calling groups_inventory to load vars for managed_node3 9704 1726776653.43550: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776653.43558: Calling all_plugins_play to load vars for managed_node3 9704 1726776653.43560: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776653.43563: Calling groups_plugins_play to load vars for managed_node3 9704 1726776653.43737: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776653.43939: done with get_vars() 9704 1726776653.43951: done getting variables 9704 1726776653.44046: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:67 Thursday 19 September 2024 16:10:53 -0400 (0:00:00.052) 0:00:06.369 **** 9704 1726776653.44068: entering _queue_task() for managed_node3/service 9704 1726776653.44070: Creating lock for service 9704 1726776653.44280: worker is 1 (out of 1 available) 9704 1726776653.44294: exiting _queue_task() for managed_node3/service 9704 1726776653.44305: done queuing things up, now waiting for results queue to drain 9704 1726776653.44307: waiting for pending results... 9898 1726776653.44430: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started 9898 1726776653.44533: in run() - task 120fa90a-8a95-fee7-0cea-000000000052 9898 1726776653.44549: variable 'ansible_search_path' from source: unknown 9898 1726776653.44553: variable 'ansible_search_path' from source: unknown 9898 1726776653.44590: variable '__kernel_settings_services' from source: include_vars 9898 1726776653.44877: variable '__kernel_settings_services' from source: include_vars 9898 1726776653.44936: variable 'omit' from source: magic vars 9898 1726776653.45009: variable 'ansible_host' from source: host vars for 'managed_node3' 9898 1726776653.45020: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9898 1726776653.45032: variable 'omit' from source: magic vars 9898 1726776653.45086: variable 'omit' from source: magic vars 9898 1726776653.45120: variable 'omit' from source: magic vars 9898 1726776653.45152: variable 'item' from source: unknown 9898 1726776653.45211: variable 'item' from source: unknown 9898 1726776653.45234: variable 'omit' from source: magic vars 9898 1726776653.45266: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9898 1726776653.45292: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9898 1726776653.45312: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9898 1726776653.45327: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9898 1726776653.45338: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9898 1726776653.45361: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9898 1726776653.45365: variable 'ansible_host' from source: host vars for 'managed_node3' 9898 1726776653.45369: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9898 1726776653.45441: Set connection var ansible_module_compression to ZIP_DEFLATED 9898 1726776653.45448: Set connection var ansible_shell_type to sh 9898 1726776653.45456: Set connection var ansible_timeout to 10 9898 1726776653.45462: Set connection var ansible_pipelining to False 9898 1726776653.45469: Set connection var ansible_shell_executable to /bin/sh 9898 1726776653.45471: Set connection var ansible_connection to ssh 9898 1726776653.45485: variable 'ansible_shell_executable' from source: unknown 9898 1726776653.45488: variable 'ansible_connection' from source: unknown 9898 1726776653.45491: variable 'ansible_module_compression' from source: unknown 9898 1726776653.45495: variable 'ansible_shell_type' from source: unknown 9898 1726776653.45498: variable 'ansible_shell_executable' from source: unknown 9898 1726776653.45501: variable 'ansible_host' from source: host vars for 'managed_node3' 9898 1726776653.45506: variable 'ansible_pipelining' from source: unknown 9898 1726776653.45509: variable 'ansible_timeout' from source: unknown 9898 1726776653.45512: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9898 1726776653.45604: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9898 1726776653.45616: variable 'omit' from source: magic vars 9898 1726776653.45623: starting attempt loop 9898 1726776653.45626: running the handler 9898 1726776653.45692: variable 'ansible_facts' from source: unknown 9898 1726776653.45774: _low_level_execute_command(): starting 9898 1726776653.45782: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9898 1726776653.48103: stdout chunk (state=2): >>>/root <<< 9898 1726776653.48224: stderr chunk (state=3): >>><<< 9898 1726776653.48233: stdout chunk (state=3): >>><<< 9898 1726776653.48252: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9898 1726776653.48267: _low_level_execute_command(): starting 9898 1726776653.48273: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771 `" && echo ansible-tmp-1726776653.4826114-9898-85141519098771="` echo /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771 `" ) && sleep 0' 9898 1726776653.50680: stdout chunk (state=2): >>>ansible-tmp-1726776653.4826114-9898-85141519098771=/root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771 <<< 9898 1726776653.50804: stderr chunk (state=3): >>><<< 9898 1726776653.50811: stdout chunk (state=3): >>><<< 9898 1726776653.50826: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776653.4826114-9898-85141519098771=/root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771 , stderr= 9898 1726776653.50854: variable 'ansible_module_compression' from source: unknown 9898 1726776653.50897: ANSIBALLZ: Using generic lock for ansible.legacy.systemd 9898 1726776653.50902: ANSIBALLZ: Acquiring lock 9898 1726776653.50905: ANSIBALLZ: Lock acquired: 140007686719920 9898 1726776653.50909: ANSIBALLZ: Creating module 9898 1726776653.72497: ANSIBALLZ: Writing module into payload 9898 1726776653.72711: ANSIBALLZ: Writing module 9898 1726776653.72758: ANSIBALLZ: Renaming module 9898 1726776653.72767: ANSIBALLZ: Done creating module 9898 1726776653.72804: variable 'ansible_facts' from source: unknown 9898 1726776653.72968: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/AnsiballZ_systemd.py 9898 1726776653.73406: Sending initial data 9898 1726776653.73412: Sent initial data (153 bytes) 9898 1726776653.75933: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp1vrly0zn /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/AnsiballZ_systemd.py <<< 9898 1726776653.77697: stderr chunk (state=3): >>><<< 9898 1726776653.77704: stdout chunk (state=3): >>><<< 9898 1726776653.77723: done transferring module to remote 9898 1726776653.77735: _low_level_execute_command(): starting 9898 1726776653.77741: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/ /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/AnsiballZ_systemd.py && sleep 0' 9898 1726776653.80100: stderr chunk (state=2): >>><<< 9898 1726776653.80107: stdout chunk (state=2): >>><<< 9898 1726776653.80120: _low_level_execute_command() done: rc=0, stdout=, stderr= 9898 1726776653.80124: _low_level_execute_command(): starting 9898 1726776653.80131: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/AnsiballZ_systemd.py && sleep 0' 9898 1726776654.08061: stdout chunk (state=2): >>> {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:31 EDT", "WatchdogTimestampMonotonic": "239516555", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "9800", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ExecMainStartTimestampMonotonic": "239379769", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9800", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:31 EDT] ; stop_time=[n/a] ; pid=9800 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15011840", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryH<<< 9898 1726776654.08123: stdout chunk (state=3): >>>igh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:31 EDT", "StateChangeTimestampMonotonic": "239516558", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveExitTimestampMonotonic": "239379946", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveEnterTimestampMonotonic": "239516558", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveExitTimestampMonotonic": "239266126", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveEnterTimestampMonotonic": "239376765", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ConditionTimestampMonotonic": "239377966", "AssertTimestamp": "Thu 2024-09-19 16:10:31 EDT", "AssertTimestampMonotonic": "239377967", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "d22f365eaba74b3a87958cacc5d42cbe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 9898 1726776654.09799: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9898 1726776654.09847: stderr chunk (state=3): >>><<< 9898 1726776654.09854: stdout chunk (state=3): >>><<< 9898 1726776654.09873: _low_level_execute_command() done: rc=0, stdout= {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:31 EDT", "WatchdogTimestampMonotonic": "239516555", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "9800", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ExecMainStartTimestampMonotonic": "239379769", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9800", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:31 EDT] ; stop_time=[n/a] ; pid=9800 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15011840", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:31 EDT", "StateChangeTimestampMonotonic": "239516558", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveExitTimestampMonotonic": "239379946", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveEnterTimestampMonotonic": "239516558", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveExitTimestampMonotonic": "239266126", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveEnterTimestampMonotonic": "239376765", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ConditionTimestampMonotonic": "239377966", "AssertTimestamp": "Thu 2024-09-19 16:10:31 EDT", "AssertTimestampMonotonic": "239377967", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "d22f365eaba74b3a87958cacc5d42cbe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=Shared connection to 10.31.8.186 closed. 9898 1726776654.09976: done with _execute_module (ansible.legacy.systemd, {'name': 'tuned', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9898 1726776654.09993: _low_level_execute_command(): starting 9898 1726776654.09998: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776653.4826114-9898-85141519098771/ > /dev/null 2>&1 && sleep 0' 9898 1726776654.12373: stderr chunk (state=2): >>><<< 9898 1726776654.12381: stdout chunk (state=2): >>><<< 9898 1726776654.12398: _low_level_execute_command() done: rc=0, stdout=, stderr= 9898 1726776654.12406: handler run complete 9898 1726776654.12443: attempt loop complete, returning result 9898 1726776654.12461: variable 'item' from source: unknown 9898 1726776654.12523: variable 'item' from source: unknown ok: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": false, "enabled": true, "item": "tuned", "name": "tuned", "state": "started", "status": { "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveEnterTimestampMonotonic": "239516558", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveExitTimestampMonotonic": "239266126", "ActiveState": "active", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Thu 2024-09-19 16:10:31 EDT", "AssertTimestampMonotonic": "239377967", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "com.redhat.tuned", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ConditionTimestampMonotonic": "239377966", "ConfigurationDirectoryMode": "0755", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "ControlGroup": "/system.slice/tuned.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Dynamic System Tuning Daemon", "DevicePolicy": "auto", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9800", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ExecMainStartTimestampMonotonic": "239379769", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:31 EDT] ; stop_time=[n/a] ; pid=9800 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "tuned.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveEnterTimestampMonotonic": "239376765", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveExitTimestampMonotonic": "239379946", "InvocationID": "d22f365eaba74b3a87958cacc5d42cbe", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "9800", "MemoryAccounting": "yes", "MemoryCurrent": "15011840", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "tuned.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/tuned/tuned.pid", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2024-09-19 16:10:31 EDT", "StateChangeTimestampMonotonic": "239516558", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Thu 2024-09-19 16:10:31 EDT", "WatchdogTimestampMonotonic": "239516555", "WatchdogUSec": "0" } } 9898 1726776654.12626: dumping result to json 9898 1726776654.12648: done dumping result, returning 9898 1726776654.12656: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started [120fa90a-8a95-fee7-0cea-000000000052] 9898 1726776654.12662: sending task result for task 120fa90a-8a95-fee7-0cea-000000000052 9898 1726776654.12770: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000052 9898 1726776654.12775: WORKER PROCESS EXITING 9704 1726776654.13127: no more pending results, returning what we have 9704 1726776654.13132: results queue empty 9704 1726776654.13133: checking for any_errors_fatal 9704 1726776654.13136: done checking for any_errors_fatal 9704 1726776654.13136: checking for max_fail_percentage 9704 1726776654.13137: done checking for max_fail_percentage 9704 1726776654.13138: checking to see if all hosts have failed and the running result is not ok 9704 1726776654.13138: done checking to see if all hosts have failed 9704 1726776654.13138: getting the remaining hosts for this loop 9704 1726776654.13139: done getting the remaining hosts for this loop 9704 1726776654.13141: getting the next task for host managed_node3 9704 1726776654.13145: done getting next task for host managed_node3 9704 1726776654.13147: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists 9704 1726776654.13148: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776654.13154: getting variables 9704 1726776654.13155: in VariableManager get_vars() 9704 1726776654.13174: Calling all_inventory to load vars for managed_node3 9704 1726776654.13176: Calling groups_inventory to load vars for managed_node3 9704 1726776654.13177: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776654.13183: Calling all_plugins_play to load vars for managed_node3 9704 1726776654.13185: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776654.13186: Calling groups_plugins_play to load vars for managed_node3 9704 1726776654.13293: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776654.13408: done with get_vars() 9704 1726776654.13418: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:74 Thursday 19 September 2024 16:10:54 -0400 (0:00:00.694) 0:00:07.063 **** 9704 1726776654.13483: entering _queue_task() for managed_node3/file 9704 1726776654.13658: worker is 1 (out of 1 available) 9704 1726776654.13673: exiting _queue_task() for managed_node3/file 9704 1726776654.13684: done queuing things up, now waiting for results queue to drain 9704 1726776654.13686: waiting for pending results... 9923 1726776654.13793: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists 9923 1726776654.13894: in run() - task 120fa90a-8a95-fee7-0cea-000000000053 9923 1726776654.13909: variable 'ansible_search_path' from source: unknown 9923 1726776654.13913: variable 'ansible_search_path' from source: unknown 9923 1726776654.13943: calling self._execute() 9923 1726776654.14000: variable 'ansible_host' from source: host vars for 'managed_node3' 9923 1726776654.14010: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9923 1726776654.14018: variable 'omit' from source: magic vars 9923 1726776654.14092: variable 'omit' from source: magic vars 9923 1726776654.14125: variable 'omit' from source: magic vars 9923 1726776654.14149: variable '__kernel_settings_profile_dir' from source: role '' all vars 9923 1726776654.14357: variable '__kernel_settings_profile_dir' from source: role '' all vars 9923 1726776654.14427: variable '__kernel_settings_profile_parent' from source: set_fact 9923 1726776654.14438: variable '__kernel_settings_tuned_profile' from source: role '' all vars 9923 1726776654.14470: variable 'omit' from source: magic vars 9923 1726776654.14503: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9923 1726776654.14531: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9923 1726776654.14550: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9923 1726776654.14563: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9923 1726776654.14574: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9923 1726776654.14598: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9923 1726776654.14603: variable 'ansible_host' from source: host vars for 'managed_node3' 9923 1726776654.14607: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9923 1726776654.14674: Set connection var ansible_module_compression to ZIP_DEFLATED 9923 1726776654.14682: Set connection var ansible_shell_type to sh 9923 1726776654.14690: Set connection var ansible_timeout to 10 9923 1726776654.14695: Set connection var ansible_pipelining to False 9923 1726776654.14703: Set connection var ansible_shell_executable to /bin/sh 9923 1726776654.14706: Set connection var ansible_connection to ssh 9923 1726776654.14722: variable 'ansible_shell_executable' from source: unknown 9923 1726776654.14726: variable 'ansible_connection' from source: unknown 9923 1726776654.14731: variable 'ansible_module_compression' from source: unknown 9923 1726776654.14735: variable 'ansible_shell_type' from source: unknown 9923 1726776654.14738: variable 'ansible_shell_executable' from source: unknown 9923 1726776654.14741: variable 'ansible_host' from source: host vars for 'managed_node3' 9923 1726776654.14745: variable 'ansible_pipelining' from source: unknown 9923 1726776654.14748: variable 'ansible_timeout' from source: unknown 9923 1726776654.14753: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9923 1726776654.14887: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 9923 1726776654.14898: variable 'omit' from source: magic vars 9923 1726776654.14903: starting attempt loop 9923 1726776654.14907: running the handler 9923 1726776654.14919: _low_level_execute_command(): starting 9923 1726776654.14927: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9923 1726776654.17242: stdout chunk (state=2): >>>/root <<< 9923 1726776654.17362: stderr chunk (state=3): >>><<< 9923 1726776654.17368: stdout chunk (state=3): >>><<< 9923 1726776654.17387: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9923 1726776654.17400: _low_level_execute_command(): starting 9923 1726776654.17405: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872 `" && echo ansible-tmp-1726776654.173943-9923-227702114798872="` echo /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872 `" ) && sleep 0' 9923 1726776654.19830: stdout chunk (state=2): >>>ansible-tmp-1726776654.173943-9923-227702114798872=/root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872 <<< 9923 1726776654.19956: stderr chunk (state=3): >>><<< 9923 1726776654.19962: stdout chunk (state=3): >>><<< 9923 1726776654.19974: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776654.173943-9923-227702114798872=/root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872 , stderr= 9923 1726776654.20005: variable 'ansible_module_compression' from source: unknown 9923 1726776654.20047: ANSIBALLZ: Using lock for file 9923 1726776654.20052: ANSIBALLZ: Acquiring lock 9923 1726776654.20057: ANSIBALLZ: Lock acquired: 140007686721600 9923 1726776654.20061: ANSIBALLZ: Creating module 9923 1726776654.31147: ANSIBALLZ: Writing module into payload 9923 1726776654.31300: ANSIBALLZ: Writing module 9923 1726776654.31324: ANSIBALLZ: Renaming module 9923 1726776654.31334: ANSIBALLZ: Done creating module 9923 1726776654.31347: variable 'ansible_facts' from source: unknown 9923 1726776654.31402: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/AnsiballZ_file.py 9923 1726776654.31873: Sending initial data 9923 1726776654.31881: Sent initial data (150 bytes) 9923 1726776654.34398: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp505qdpa6 /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/AnsiballZ_file.py <<< 9923 1726776654.35283: stderr chunk (state=3): >>><<< 9923 1726776654.35292: stdout chunk (state=3): >>><<< 9923 1726776654.35310: done transferring module to remote 9923 1726776654.35322: _low_level_execute_command(): starting 9923 1726776654.35329: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/ /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/AnsiballZ_file.py && sleep 0' 9923 1726776654.37637: stderr chunk (state=2): >>><<< 9923 1726776654.37645: stdout chunk (state=2): >>><<< 9923 1726776654.37659: _low_level_execute_command() done: rc=0, stdout=, stderr= 9923 1726776654.37663: _low_level_execute_command(): starting 9923 1726776654.37669: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/AnsiballZ_file.py && sleep 0' 9923 1726776654.53893: stdout chunk (state=2): >>> {"path": "/etc/tuned/kernel_settings", "changed": true, "diff": {"before": {"path": "/etc/tuned/kernel_settings", "state": "absent"}, "after": {"path": "/etc/tuned/kernel_settings", "state": "directory"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0755", "state": "directory", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 6, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "directory", "mode": "0755", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 9923 1726776654.55020: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9923 1726776654.55070: stderr chunk (state=3): >>><<< 9923 1726776654.55079: stdout chunk (state=3): >>><<< 9923 1726776654.55095: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/kernel_settings", "changed": true, "diff": {"before": {"path": "/etc/tuned/kernel_settings", "state": "absent"}, "after": {"path": "/etc/tuned/kernel_settings", "state": "directory"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0755", "state": "directory", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 6, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "directory", "mode": "0755", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 9923 1726776654.55127: done with _execute_module (file, {'path': '/etc/tuned/kernel_settings', 'state': 'directory', 'mode': '0755', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9923 1726776654.55139: _low_level_execute_command(): starting 9923 1726776654.55147: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776654.173943-9923-227702114798872/ > /dev/null 2>&1 && sleep 0' 9923 1726776654.57568: stderr chunk (state=2): >>><<< 9923 1726776654.57579: stdout chunk (state=2): >>><<< 9923 1726776654.57594: _low_level_execute_command() done: rc=0, stdout=, stderr= 9923 1726776654.57601: handler run complete 9923 1726776654.57621: attempt loop complete, returning result 9923 1726776654.57625: _execute() done 9923 1726776654.57630: dumping result to json 9923 1726776654.57636: done dumping result, returning 9923 1726776654.57644: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists [120fa90a-8a95-fee7-0cea-000000000053] 9923 1726776654.57650: sending task result for task 120fa90a-8a95-fee7-0cea-000000000053 9923 1726776654.57687: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000053 9923 1726776654.57691: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/tuned/kernel_settings", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 6, "state": "directory", "uid": 0 } 9704 1726776654.57845: no more pending results, returning what we have 9704 1726776654.57847: results queue empty 9704 1726776654.57848: checking for any_errors_fatal 9704 1726776654.57860: done checking for any_errors_fatal 9704 1726776654.57861: checking for max_fail_percentage 9704 1726776654.57863: done checking for max_fail_percentage 9704 1726776654.57863: checking to see if all hosts have failed and the running result is not ok 9704 1726776654.57864: done checking to see if all hosts have failed 9704 1726776654.57865: getting the remaining hosts for this loop 9704 1726776654.57866: done getting the remaining hosts for this loop 9704 1726776654.57869: getting the next task for host managed_node3 9704 1726776654.57873: done getting next task for host managed_node3 9704 1726776654.57876: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get active_profile 9704 1726776654.57878: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776654.57886: getting variables 9704 1726776654.57888: in VariableManager get_vars() 9704 1726776654.57920: Calling all_inventory to load vars for managed_node3 9704 1726776654.57922: Calling groups_inventory to load vars for managed_node3 9704 1726776654.57924: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776654.57934: Calling all_plugins_play to load vars for managed_node3 9704 1726776654.57937: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776654.57939: Calling groups_plugins_play to load vars for managed_node3 9704 1726776654.58076: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776654.58196: done with get_vars() 9704 1726776654.58204: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Get active_profile] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:80 Thursday 19 September 2024 16:10:54 -0400 (0:00:00.447) 0:00:07.511 **** 9704 1726776654.58271: entering _queue_task() for managed_node3/slurp 9704 1726776654.58445: worker is 1 (out of 1 available) 9704 1726776654.58459: exiting _queue_task() for managed_node3/slurp 9704 1726776654.58470: done queuing things up, now waiting for results queue to drain 9704 1726776654.58473: waiting for pending results... 9939 1726776654.58594: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get active_profile 9939 1726776654.58689: in run() - task 120fa90a-8a95-fee7-0cea-000000000054 9939 1726776654.58705: variable 'ansible_search_path' from source: unknown 9939 1726776654.58709: variable 'ansible_search_path' from source: unknown 9939 1726776654.58739: calling self._execute() 9939 1726776654.58794: variable 'ansible_host' from source: host vars for 'managed_node3' 9939 1726776654.58803: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9939 1726776654.58811: variable 'omit' from source: magic vars 9939 1726776654.58884: variable 'omit' from source: magic vars 9939 1726776654.58913: variable 'omit' from source: magic vars 9939 1726776654.58938: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 9939 1726776654.59145: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 9939 1726776654.59203: variable '__kernel_settings_tuned_dir' from source: role '' all vars 9939 1726776654.59234: variable 'omit' from source: magic vars 9939 1726776654.59266: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9939 1726776654.59292: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9939 1726776654.59310: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9939 1726776654.59325: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9939 1726776654.59336: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9939 1726776654.59360: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9939 1726776654.59365: variable 'ansible_host' from source: host vars for 'managed_node3' 9939 1726776654.59369: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9939 1726776654.59440: Set connection var ansible_module_compression to ZIP_DEFLATED 9939 1726776654.59448: Set connection var ansible_shell_type to sh 9939 1726776654.59456: Set connection var ansible_timeout to 10 9939 1726776654.59461: Set connection var ansible_pipelining to False 9939 1726776654.59469: Set connection var ansible_shell_executable to /bin/sh 9939 1726776654.59472: Set connection var ansible_connection to ssh 9939 1726776654.59489: variable 'ansible_shell_executable' from source: unknown 9939 1726776654.59493: variable 'ansible_connection' from source: unknown 9939 1726776654.59496: variable 'ansible_module_compression' from source: unknown 9939 1726776654.59497: variable 'ansible_shell_type' from source: unknown 9939 1726776654.59499: variable 'ansible_shell_executable' from source: unknown 9939 1726776654.59501: variable 'ansible_host' from source: host vars for 'managed_node3' 9939 1726776654.59503: variable 'ansible_pipelining' from source: unknown 9939 1726776654.59504: variable 'ansible_timeout' from source: unknown 9939 1726776654.59506: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9939 1726776654.59652: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 9939 1726776654.59663: variable 'omit' from source: magic vars 9939 1726776654.59668: starting attempt loop 9939 1726776654.59672: running the handler 9939 1726776654.59683: _low_level_execute_command(): starting 9939 1726776654.59691: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9939 1726776654.62179: stdout chunk (state=2): >>>/root <<< 9939 1726776654.62294: stderr chunk (state=3): >>><<< 9939 1726776654.62300: stdout chunk (state=3): >>><<< 9939 1726776654.62316: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9939 1726776654.62335: _low_level_execute_command(): starting 9939 1726776654.62342: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544 `" && echo ansible-tmp-1726776654.6232576-9939-156151950414544="` echo /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544 `" ) && sleep 0' 9939 1726776654.65067: stdout chunk (state=2): >>>ansible-tmp-1726776654.6232576-9939-156151950414544=/root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544 <<< 9939 1726776654.65109: stderr chunk (state=3): >>><<< 9939 1726776654.65120: stdout chunk (state=3): >>><<< 9939 1726776654.65141: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776654.6232576-9939-156151950414544=/root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544 , stderr= 9939 1726776654.65191: variable 'ansible_module_compression' from source: unknown 9939 1726776654.65241: ANSIBALLZ: Using lock for slurp 9939 1726776654.65247: ANSIBALLZ: Acquiring lock 9939 1726776654.65251: ANSIBALLZ: Lock acquired: 140007686720544 9939 1726776654.65254: ANSIBALLZ: Creating module 9939 1726776654.76775: ANSIBALLZ: Writing module into payload 9939 1726776654.76852: ANSIBALLZ: Writing module 9939 1726776654.76877: ANSIBALLZ: Renaming module 9939 1726776654.76887: ANSIBALLZ: Done creating module 9939 1726776654.76904: variable 'ansible_facts' from source: unknown 9939 1726776654.76990: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/AnsiballZ_slurp.py 9939 1726776654.77486: Sending initial data 9939 1726776654.77493: Sent initial data (152 bytes) 9939 1726776654.80020: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpdwxx_c56 /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/AnsiballZ_slurp.py <<< 9939 1726776654.80982: stderr chunk (state=3): >>><<< 9939 1726776654.80990: stdout chunk (state=3): >>><<< 9939 1726776654.81007: done transferring module to remote 9939 1726776654.81017: _low_level_execute_command(): starting 9939 1726776654.81023: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/ /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/AnsiballZ_slurp.py && sleep 0' 9939 1726776654.83479: stderr chunk (state=2): >>><<< 9939 1726776654.83486: stdout chunk (state=2): >>><<< 9939 1726776654.83499: _low_level_execute_command() done: rc=0, stdout=, stderr= 9939 1726776654.83504: _low_level_execute_command(): starting 9939 1726776654.83509: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/AnsiballZ_slurp.py && sleep 0' 9939 1726776654.98404: stdout chunk (state=2): >>> {"content": "dmlydHVhbC1ndWVzdAo=", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} <<< 9939 1726776654.99453: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9939 1726776654.99503: stderr chunk (state=3): >>><<< 9939 1726776654.99510: stdout chunk (state=3): >>><<< 9939 1726776654.99528: _low_level_execute_command() done: rc=0, stdout= {"content": "dmlydHVhbC1ndWVzdAo=", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} , stderr=Shared connection to 10.31.8.186 closed. 9939 1726776654.99554: done with _execute_module (slurp, {'path': '/etc/tuned/active_profile', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9939 1726776654.99564: _low_level_execute_command(): starting 9939 1726776654.99570: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776654.6232576-9939-156151950414544/ > /dev/null 2>&1 && sleep 0' 9939 1726776655.02036: stderr chunk (state=2): >>><<< 9939 1726776655.02043: stdout chunk (state=2): >>><<< 9939 1726776655.02057: _low_level_execute_command() done: rc=0, stdout=, stderr= 9939 1726776655.02064: handler run complete 9939 1726776655.02081: attempt loop complete, returning result 9939 1726776655.02086: _execute() done 9939 1726776655.02090: dumping result to json 9939 1726776655.02093: done dumping result, returning 9939 1726776655.02100: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get active_profile [120fa90a-8a95-fee7-0cea-000000000054] 9939 1726776655.02104: sending task result for task 120fa90a-8a95-fee7-0cea-000000000054 9939 1726776655.02143: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000054 9939 1726776655.02146: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "dmlydHVhbC1ndWVzdAo=", "encoding": "base64", "source": "/etc/tuned/active_profile" } 9704 1726776655.02657: no more pending results, returning what we have 9704 1726776655.02661: results queue empty 9704 1726776655.02662: checking for any_errors_fatal 9704 1726776655.02671: done checking for any_errors_fatal 9704 1726776655.02671: checking for max_fail_percentage 9704 1726776655.02673: done checking for max_fail_percentage 9704 1726776655.02674: checking to see if all hosts have failed and the running result is not ok 9704 1726776655.02674: done checking to see if all hosts have failed 9704 1726776655.02675: getting the remaining hosts for this loop 9704 1726776655.02676: done getting the remaining hosts for this loop 9704 1726776655.02679: getting the next task for host managed_node3 9704 1726776655.02684: done getting next task for host managed_node3 9704 1726776655.02688: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set active_profile 9704 1726776655.02690: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776655.02703: getting variables 9704 1726776655.02704: in VariableManager get_vars() 9704 1726776655.02741: Calling all_inventory to load vars for managed_node3 9704 1726776655.02744: Calling groups_inventory to load vars for managed_node3 9704 1726776655.02746: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776655.02754: Calling all_plugins_play to load vars for managed_node3 9704 1726776655.02757: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776655.02759: Calling groups_plugins_play to load vars for managed_node3 9704 1726776655.02924: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776655.03112: done with get_vars() 9704 1726776655.03122: done getting variables 9704 1726776655.03164: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set active_profile] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:85 Thursday 19 September 2024 16:10:55 -0400 (0:00:00.449) 0:00:07.960 **** 9704 1726776655.03185: entering _queue_task() for managed_node3/set_fact 9704 1726776655.03357: worker is 1 (out of 1 available) 9704 1726776655.03370: exiting _queue_task() for managed_node3/set_fact 9704 1726776655.03382: done queuing things up, now waiting for results queue to drain 9704 1726776655.03384: waiting for pending results... 9960 1726776655.03494: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set active_profile 9960 1726776655.03593: in run() - task 120fa90a-8a95-fee7-0cea-000000000055 9960 1726776655.03609: variable 'ansible_search_path' from source: unknown 9960 1726776655.03613: variable 'ansible_search_path' from source: unknown 9960 1726776655.03644: calling self._execute() 9960 1726776655.03772: variable 'ansible_host' from source: host vars for 'managed_node3' 9960 1726776655.03781: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9960 1726776655.03790: variable 'omit' from source: magic vars 9960 1726776655.03862: variable 'omit' from source: magic vars 9960 1726776655.03896: variable 'omit' from source: magic vars 9960 1726776655.04166: variable '__kernel_settings_tuned_profile' from source: role '' all vars 9960 1726776655.04177: variable '__cur_profile' from source: task vars 9960 1726776655.04278: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 9960 1726776655.06165: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 9960 1726776655.06224: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 9960 1726776655.06278: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 9960 1726776655.06315: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 9960 1726776655.06342: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 9960 1726776655.06415: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 9960 1726776655.06440: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 9960 1726776655.06469: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 9960 1726776655.06498: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 9960 1726776655.06518: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 9960 1726776655.06623: variable '__kernel_settings_tuned_current_profile' from source: set_fact 9960 1726776655.06662: variable 'omit' from source: magic vars 9960 1726776655.06682: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9960 1726776655.06701: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9960 1726776655.06716: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9960 1726776655.06730: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9960 1726776655.06739: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9960 1726776655.06766: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9960 1726776655.06770: variable 'ansible_host' from source: host vars for 'managed_node3' 9960 1726776655.06774: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9960 1726776655.06858: Set connection var ansible_module_compression to ZIP_DEFLATED 9960 1726776655.06866: Set connection var ansible_shell_type to sh 9960 1726776655.06874: Set connection var ansible_timeout to 10 9960 1726776655.06879: Set connection var ansible_pipelining to False 9960 1726776655.06886: Set connection var ansible_shell_executable to /bin/sh 9960 1726776655.06889: Set connection var ansible_connection to ssh 9960 1726776655.06907: variable 'ansible_shell_executable' from source: unknown 9960 1726776655.06912: variable 'ansible_connection' from source: unknown 9960 1726776655.06915: variable 'ansible_module_compression' from source: unknown 9960 1726776655.06918: variable 'ansible_shell_type' from source: unknown 9960 1726776655.06920: variable 'ansible_shell_executable' from source: unknown 9960 1726776655.06923: variable 'ansible_host' from source: host vars for 'managed_node3' 9960 1726776655.06927: variable 'ansible_pipelining' from source: unknown 9960 1726776655.06932: variable 'ansible_timeout' from source: unknown 9960 1726776655.06936: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9960 1726776655.07012: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9960 1726776655.07022: variable 'omit' from source: magic vars 9960 1726776655.07026: starting attempt loop 9960 1726776655.07030: running the handler 9960 1726776655.07041: handler run complete 9960 1726776655.07049: attempt loop complete, returning result 9960 1726776655.07053: _execute() done 9960 1726776655.07056: dumping result to json 9960 1726776655.07059: done dumping result, returning 9960 1726776655.07065: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set active_profile [120fa90a-8a95-fee7-0cea-000000000055] 9960 1726776655.07070: sending task result for task 120fa90a-8a95-fee7-0cea-000000000055 9960 1726776655.07091: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000055 9960 1726776655.07094: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_active_profile": "virtual-guest kernel_settings" }, "changed": false } 9704 1726776655.07290: no more pending results, returning what we have 9704 1726776655.07292: results queue empty 9704 1726776655.07293: checking for any_errors_fatal 9704 1726776655.07296: done checking for any_errors_fatal 9704 1726776655.07297: checking for max_fail_percentage 9704 1726776655.07298: done checking for max_fail_percentage 9704 1726776655.07299: checking to see if all hosts have failed and the running result is not ok 9704 1726776655.07299: done checking to see if all hosts have failed 9704 1726776655.07300: getting the remaining hosts for this loop 9704 1726776655.07301: done getting the remaining hosts for this loop 9704 1726776655.07304: getting the next task for host managed_node3 9704 1726776655.07308: done getting next task for host managed_node3 9704 1726776655.07311: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile 9704 1726776655.07313: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776655.07330: getting variables 9704 1726776655.07331: in VariableManager get_vars() 9704 1726776655.07356: Calling all_inventory to load vars for managed_node3 9704 1726776655.07359: Calling groups_inventory to load vars for managed_node3 9704 1726776655.07361: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776655.07369: Calling all_plugins_play to load vars for managed_node3 9704 1726776655.07372: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776655.07376: Calling groups_plugins_play to load vars for managed_node3 9704 1726776655.07531: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776655.07724: done with get_vars() 9704 1726776655.07735: done getting variables 9704 1726776655.07848: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:91 Thursday 19 September 2024 16:10:55 -0400 (0:00:00.046) 0:00:08.007 **** 9704 1726776655.07878: entering _queue_task() for managed_node3/copy 9704 1726776655.08043: worker is 1 (out of 1 available) 9704 1726776655.08057: exiting _queue_task() for managed_node3/copy 9704 1726776655.08067: done queuing things up, now waiting for results queue to drain 9704 1726776655.08069: waiting for pending results... 9962 1726776655.08183: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile 9962 1726776655.08284: in run() - task 120fa90a-8a95-fee7-0cea-000000000056 9962 1726776655.08299: variable 'ansible_search_path' from source: unknown 9962 1726776655.08302: variable 'ansible_search_path' from source: unknown 9962 1726776655.08327: calling self._execute() 9962 1726776655.08385: variable 'ansible_host' from source: host vars for 'managed_node3' 9962 1726776655.08391: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9962 1726776655.08398: variable 'omit' from source: magic vars 9962 1726776655.08470: variable 'omit' from source: magic vars 9962 1726776655.08501: variable 'omit' from source: magic vars 9962 1726776655.08520: variable '__kernel_settings_active_profile' from source: set_fact 9962 1726776655.08732: variable '__kernel_settings_active_profile' from source: set_fact 9962 1726776655.08751: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 9962 1726776655.08800: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 9962 1726776655.08853: variable '__kernel_settings_tuned_dir' from source: role '' all vars 9962 1726776655.08875: variable 'omit' from source: magic vars 9962 1726776655.08906: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 9962 1726776655.09171: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 9962 1726776655.09192: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 9962 1726776655.09208: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9962 1726776655.09218: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 9962 1726776655.09244: variable 'inventory_hostname' from source: host vars for 'managed_node3' 9962 1726776655.09249: variable 'ansible_host' from source: host vars for 'managed_node3' 9962 1726776655.09254: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9962 1726776655.09323: Set connection var ansible_module_compression to ZIP_DEFLATED 9962 1726776655.09333: Set connection var ansible_shell_type to sh 9962 1726776655.09341: Set connection var ansible_timeout to 10 9962 1726776655.09346: Set connection var ansible_pipelining to False 9962 1726776655.09353: Set connection var ansible_shell_executable to /bin/sh 9962 1726776655.09356: Set connection var ansible_connection to ssh 9962 1726776655.09372: variable 'ansible_shell_executable' from source: unknown 9962 1726776655.09376: variable 'ansible_connection' from source: unknown 9962 1726776655.09379: variable 'ansible_module_compression' from source: unknown 9962 1726776655.09383: variable 'ansible_shell_type' from source: unknown 9962 1726776655.09386: variable 'ansible_shell_executable' from source: unknown 9962 1726776655.09389: variable 'ansible_host' from source: host vars for 'managed_node3' 9962 1726776655.09393: variable 'ansible_pipelining' from source: unknown 9962 1726776655.09396: variable 'ansible_timeout' from source: unknown 9962 1726776655.09400: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 9962 1726776655.09488: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 9962 1726776655.09500: variable 'omit' from source: magic vars 9962 1726776655.09506: starting attempt loop 9962 1726776655.09510: running the handler 9962 1726776655.09521: _low_level_execute_command(): starting 9962 1726776655.09531: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 9962 1726776655.11958: stdout chunk (state=2): >>>/root <<< 9962 1726776655.12083: stderr chunk (state=3): >>><<< 9962 1726776655.12091: stdout chunk (state=3): >>><<< 9962 1726776655.12111: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 9962 1726776655.12129: _low_level_execute_command(): starting 9962 1726776655.12139: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982 `" && echo ansible-tmp-1726776655.1211996-9962-208842373916982="` echo /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982 `" ) && sleep 0' 9962 1726776655.14611: stdout chunk (state=2): >>>ansible-tmp-1726776655.1211996-9962-208842373916982=/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982 <<< 9962 1726776655.14741: stderr chunk (state=3): >>><<< 9962 1726776655.14748: stdout chunk (state=3): >>><<< 9962 1726776655.14763: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776655.1211996-9962-208842373916982=/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982 , stderr= 9962 1726776655.14839: variable 'ansible_module_compression' from source: unknown 9962 1726776655.14881: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 9962 1726776655.14910: variable 'ansible_facts' from source: unknown 9962 1726776655.14978: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_stat.py 9962 1726776655.15063: Sending initial data 9962 1726776655.15070: Sent initial data (151 bytes) 9962 1726776655.17575: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpd5g9uwlp /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_stat.py <<< 9962 1726776655.18559: stderr chunk (state=3): >>><<< 9962 1726776655.18566: stdout chunk (state=3): >>><<< 9962 1726776655.18586: done transferring module to remote 9962 1726776655.18596: _low_level_execute_command(): starting 9962 1726776655.18601: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/ /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_stat.py && sleep 0' 9962 1726776655.20936: stderr chunk (state=2): >>><<< 9962 1726776655.20944: stdout chunk (state=2): >>><<< 9962 1726776655.20958: _low_level_execute_command() done: rc=0, stdout=, stderr= 9962 1726776655.20962: _low_level_execute_command(): starting 9962 1726776655.20967: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_stat.py && sleep 0' 9962 1726776655.37444: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 14, "inode": 79692034, "dev": 51713, "nlink": 1, "atime": 1726776654.9821959, "mtime": 1726776642.7430372, "ctime": 1726776643.0060408, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "mimetype": "text/plain", "charset": "us-ascii", "version": "2860405357", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 9962 1726776655.38801: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9962 1726776655.38812: stdout chunk (state=3): >>><<< 9962 1726776655.38827: stderr chunk (state=3): >>><<< 9962 1726776655.38845: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 14, "inode": 79692034, "dev": 51713, "nlink": 1, "atime": 1726776654.9821959, "mtime": 1726776642.7430372, "ctime": 1726776643.0060408, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "mimetype": "text/plain", "charset": "us-ascii", "version": "2860405357", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 9962 1726776655.38901: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/active_profile', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9962 1726776655.39377: Sending initial data 9962 1726776655.39384: Sent initial data (140 bytes) 9962 1726776655.42098: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp77jkcgrp /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source <<< 9962 1726776655.42460: stderr chunk (state=3): >>><<< 9962 1726776655.42470: stdout chunk (state=3): >>><<< 9962 1726776655.42493: _low_level_execute_command(): starting 9962 1726776655.42504: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/ /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source && sleep 0' 9962 1726776655.45097: stderr chunk (state=2): >>><<< 9962 1726776655.45107: stdout chunk (state=2): >>><<< 9962 1726776655.45125: _low_level_execute_command() done: rc=0, stdout=, stderr= 9962 1726776655.45151: variable 'ansible_module_compression' from source: unknown 9962 1726776655.45201: ANSIBALLZ: Using generic lock for ansible.legacy.copy 9962 1726776655.45206: ANSIBALLZ: Acquiring lock 9962 1726776655.45210: ANSIBALLZ: Lock acquired: 140007686719920 9962 1726776655.45213: ANSIBALLZ: Creating module 9962 1726776655.56356: ANSIBALLZ: Writing module into payload 9962 1726776655.56491: ANSIBALLZ: Writing module 9962 1726776655.56512: ANSIBALLZ: Renaming module 9962 1726776655.56522: ANSIBALLZ: Done creating module 9962 1726776655.56535: variable 'ansible_facts' from source: unknown 9962 1726776655.56588: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_copy.py 9962 1726776655.56677: Sending initial data 9962 1726776655.56684: Sent initial data (151 bytes) 9962 1726776655.59215: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp4jptsxwd /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_copy.py <<< 9962 1726776655.60206: stderr chunk (state=3): >>><<< 9962 1726776655.60213: stdout chunk (state=3): >>><<< 9962 1726776655.60235: done transferring module to remote 9962 1726776655.60244: _low_level_execute_command(): starting 9962 1726776655.60249: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/ /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_copy.py && sleep 0' 9962 1726776655.62669: stderr chunk (state=2): >>><<< 9962 1726776655.62679: stdout chunk (state=2): >>><<< 9962 1726776655.62693: _low_level_execute_command() done: rc=0, stdout=, stderr= 9962 1726776655.62697: _low_level_execute_command(): starting 9962 1726776655.62703: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/AnsiballZ_copy.py && sleep 0' 9962 1726776655.79114: stdout chunk (state=2): >>> {"dest": "/etc/tuned/active_profile", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source", "md5sum": "8d80fe3f09ba4b9ac8d7fd5e8541a324", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "invocation": {"module_args": {"dest": "/etc/tuned/active_profile", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source", "_original_basename": "tmp77jkcgrp", "follow": false, "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 9962 1726776655.80309: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 9962 1726776655.80359: stderr chunk (state=3): >>><<< 9962 1726776655.80366: stdout chunk (state=3): >>><<< 9962 1726776655.80382: _low_level_execute_command() done: rc=0, stdout= {"dest": "/etc/tuned/active_profile", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source", "md5sum": "8d80fe3f09ba4b9ac8d7fd5e8541a324", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "invocation": {"module_args": {"dest": "/etc/tuned/active_profile", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source", "_original_basename": "tmp77jkcgrp", "follow": false, "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 9962 1726776655.80407: done with _execute_module (ansible.legacy.copy, {'dest': '/etc/tuned/active_profile', 'mode': '0600', 'src': '/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source', '_original_basename': 'tmp77jkcgrp', 'follow': False, 'checksum': 'a79569d3860cb6a066e0e92c8b22ffd0e8796bfd', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 9962 1726776655.80418: _low_level_execute_command(): starting 9962 1726776655.80424: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/ > /dev/null 2>&1 && sleep 0' 9962 1726776655.82856: stderr chunk (state=2): >>><<< 9962 1726776655.82866: stdout chunk (state=2): >>><<< 9962 1726776655.82881: _low_level_execute_command() done: rc=0, stdout=, stderr= 9962 1726776655.82890: handler run complete 9962 1726776655.82917: attempt loop complete, returning result 9962 1726776655.82925: _execute() done 9962 1726776655.82930: dumping result to json 9962 1726776655.82937: done dumping result, returning 9962 1726776655.82944: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile [120fa90a-8a95-fee7-0cea-000000000056] 9962 1726776655.82949: sending task result for task 120fa90a-8a95-fee7-0cea-000000000056 9962 1726776655.82991: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000056 9962 1726776655.82995: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "dest": "/etc/tuned/active_profile", "gid": 0, "group": "root", "md5sum": "8d80fe3f09ba4b9ac8d7fd5e8541a324", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "src": "/root/.ansible/tmp/ansible-tmp-1726776655.1211996-9962-208842373916982/source", "state": "file", "uid": 0 } 9704 1726776655.83441: no more pending results, returning what we have 9704 1726776655.83445: results queue empty 9704 1726776655.83445: checking for any_errors_fatal 9704 1726776655.83449: done checking for any_errors_fatal 9704 1726776655.83449: checking for max_fail_percentage 9704 1726776655.83450: done checking for max_fail_percentage 9704 1726776655.83450: checking to see if all hosts have failed and the running result is not ok 9704 1726776655.83451: done checking to see if all hosts have failed 9704 1726776655.83451: getting the remaining hosts for this loop 9704 1726776655.83452: done getting the remaining hosts for this loop 9704 1726776655.83455: getting the next task for host managed_node3 9704 1726776655.83458: done getting next task for host managed_node3 9704 1726776655.83461: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual 9704 1726776655.83462: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776655.83469: getting variables 9704 1726776655.83470: in VariableManager get_vars() 9704 1726776655.83493: Calling all_inventory to load vars for managed_node3 9704 1726776655.83495: Calling groups_inventory to load vars for managed_node3 9704 1726776655.83496: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776655.83502: Calling all_plugins_play to load vars for managed_node3 9704 1726776655.83504: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776655.83505: Calling groups_plugins_play to load vars for managed_node3 9704 1726776655.83783: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776655.83894: done with get_vars() 9704 1726776655.83902: done getting variables 9704 1726776655.83945: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set profile_mode to manual] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:99 Thursday 19 September 2024 16:10:55 -0400 (0:00:00.760) 0:00:08.768 **** 9704 1726776655.83965: entering _queue_task() for managed_node3/copy 9704 1726776655.84134: worker is 1 (out of 1 available) 9704 1726776655.84148: exiting _queue_task() for managed_node3/copy 9704 1726776655.84159: done queuing things up, now waiting for results queue to drain 9704 1726776655.84161: waiting for pending results... 10000 1726776655.84277: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual 10000 1726776655.84383: in run() - task 120fa90a-8a95-fee7-0cea-000000000057 10000 1726776655.84399: variable 'ansible_search_path' from source: unknown 10000 1726776655.84403: variable 'ansible_search_path' from source: unknown 10000 1726776655.84436: calling self._execute() 10000 1726776655.84495: variable 'ansible_host' from source: host vars for 'managed_node3' 10000 1726776655.84503: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10000 1726776655.84512: variable 'omit' from source: magic vars 10000 1726776655.84586: variable 'omit' from source: magic vars 10000 1726776655.84623: variable 'omit' from source: magic vars 10000 1726776655.84646: variable '__kernel_settings_tuned_profile_mode' from source: role '' all vars 10000 1726776655.84864: variable '__kernel_settings_tuned_profile_mode' from source: role '' all vars 10000 1726776655.84925: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10000 1726776655.84955: variable 'omit' from source: magic vars 10000 1726776655.84987: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10000 1726776655.85016: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10000 1726776655.85039: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10000 1726776655.85053: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10000 1726776655.85064: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10000 1726776655.85094: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10000 1726776655.85100: variable 'ansible_host' from source: host vars for 'managed_node3' 10000 1726776655.85103: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10000 1726776655.85189: Set connection var ansible_module_compression to ZIP_DEFLATED 10000 1726776655.85198: Set connection var ansible_shell_type to sh 10000 1726776655.85206: Set connection var ansible_timeout to 10 10000 1726776655.85212: Set connection var ansible_pipelining to False 10000 1726776655.85221: Set connection var ansible_shell_executable to /bin/sh 10000 1726776655.85225: Set connection var ansible_connection to ssh 10000 1726776655.85246: variable 'ansible_shell_executable' from source: unknown 10000 1726776655.85249: variable 'ansible_connection' from source: unknown 10000 1726776655.85252: variable 'ansible_module_compression' from source: unknown 10000 1726776655.85255: variable 'ansible_shell_type' from source: unknown 10000 1726776655.85256: variable 'ansible_shell_executable' from source: unknown 10000 1726776655.85258: variable 'ansible_host' from source: host vars for 'managed_node3' 10000 1726776655.85260: variable 'ansible_pipelining' from source: unknown 10000 1726776655.85262: variable 'ansible_timeout' from source: unknown 10000 1726776655.85264: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10000 1726776655.85382: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10000 1726776655.85393: variable 'omit' from source: magic vars 10000 1726776655.85399: starting attempt loop 10000 1726776655.85402: running the handler 10000 1726776655.85413: _low_level_execute_command(): starting 10000 1726776655.85422: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10000 1726776655.87889: stdout chunk (state=2): >>>/root <<< 10000 1726776655.88001: stderr chunk (state=3): >>><<< 10000 1726776655.88008: stdout chunk (state=3): >>><<< 10000 1726776655.88025: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10000 1726776655.88041: _low_level_execute_command(): starting 10000 1726776655.88047: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347 `" && echo ansible-tmp-1726776655.8803492-10000-129868350717347="` echo /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347 `" ) && sleep 0' 10000 1726776655.90470: stdout chunk (state=2): >>>ansible-tmp-1726776655.8803492-10000-129868350717347=/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347 <<< 10000 1726776655.90599: stderr chunk (state=3): >>><<< 10000 1726776655.90605: stdout chunk (state=3): >>><<< 10000 1726776655.90619: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776655.8803492-10000-129868350717347=/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347 , stderr= 10000 1726776655.90689: variable 'ansible_module_compression' from source: unknown 10000 1726776655.90737: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10000 1726776655.90767: variable 'ansible_facts' from source: unknown 10000 1726776655.90836: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_stat.py 10000 1726776655.90918: Sending initial data 10000 1726776655.90927: Sent initial data (152 bytes) 10000 1726776655.93364: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpzvxok4im /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_stat.py <<< 10000 1726776655.94338: stderr chunk (state=3): >>><<< 10000 1726776655.94345: stdout chunk (state=3): >>><<< 10000 1726776655.94364: done transferring module to remote 10000 1726776655.94375: _low_level_execute_command(): starting 10000 1726776655.94380: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/ /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_stat.py && sleep 0' 10000 1726776655.96710: stderr chunk (state=2): >>><<< 10000 1726776655.96718: stdout chunk (state=2): >>><<< 10000 1726776655.96733: _low_level_execute_command() done: rc=0, stdout=, stderr= 10000 1726776655.96737: _low_level_execute_command(): starting 10000 1726776655.96743: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_stat.py && sleep 0' 10000 1726776656.12618: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 5, "inode": 85983490, "dev": 51713, "nlink": 1, "atime": 1726776643.634049, "mtime": 1726776643.3880455, "ctime": 1726776643.6350489, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "mimetype": "text/plain", "charset": "us-ascii", "version": "3218256690", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10000 1726776656.13761: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10000 1726776656.13801: stderr chunk (state=3): >>><<< 10000 1726776656.13810: stdout chunk (state=3): >>><<< 10000 1726776656.13825: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 5, "inode": 85983490, "dev": 51713, "nlink": 1, "atime": 1726776643.634049, "mtime": 1726776643.3880455, "ctime": 1726776643.6350489, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "mimetype": "text/plain", "charset": "us-ascii", "version": "3218256690", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10000 1726776656.13868: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/profile_mode', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10000 1726776656.13950: Sending initial data 10000 1726776656.13958: Sent initial data (141 bytes) 10000 1726776656.16432: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpm3xpwwsg /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source <<< 10000 1726776656.16672: stderr chunk (state=3): >>><<< 10000 1726776656.16679: stdout chunk (state=3): >>><<< 10000 1726776656.16696: _low_level_execute_command(): starting 10000 1726776656.16704: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/ /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source && sleep 0' 10000 1726776656.18990: stderr chunk (state=2): >>><<< 10000 1726776656.18997: stdout chunk (state=2): >>><<< 10000 1726776656.19011: _low_level_execute_command() done: rc=0, stdout=, stderr= 10000 1726776656.19038: variable 'ansible_module_compression' from source: unknown 10000 1726776656.19085: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.copy-ZIP_DEFLATED 10000 1726776656.19106: variable 'ansible_facts' from source: unknown 10000 1726776656.19197: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_copy.py 10000 1726776656.19597: Sending initial data 10000 1726776656.19603: Sent initial data (152 bytes) 10000 1726776656.21813: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp4ru8c24w /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_copy.py <<< 10000 1726776656.22879: stderr chunk (state=3): >>><<< 10000 1726776656.22885: stdout chunk (state=3): >>><<< 10000 1726776656.22903: done transferring module to remote 10000 1726776656.22911: _low_level_execute_command(): starting 10000 1726776656.22916: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/ /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_copy.py && sleep 0' 10000 1726776656.25177: stderr chunk (state=2): >>><<< 10000 1726776656.25185: stdout chunk (state=2): >>><<< 10000 1726776656.25197: _low_level_execute_command() done: rc=0, stdout=, stderr= 10000 1726776656.25201: _low_level_execute_command(): starting 10000 1726776656.25206: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/AnsiballZ_copy.py && sleep 0' 10000 1726776656.41586: stdout chunk (state=2): >>> {"dest": "/etc/tuned/profile_mode", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source", "md5sum": "cf3f2a865fbea819dadd439586eaee31", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "invocation": {"module_args": {"dest": "/etc/tuned/profile_mode", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source", "_original_basename": "tmpm3xpwwsg", "follow": false, "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10000 1726776656.42887: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10000 1726776656.42897: stdout chunk (state=3): >>><<< 10000 1726776656.42906: stderr chunk (state=3): >>><<< 10000 1726776656.42922: _low_level_execute_command() done: rc=0, stdout= {"dest": "/etc/tuned/profile_mode", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source", "md5sum": "cf3f2a865fbea819dadd439586eaee31", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "invocation": {"module_args": {"dest": "/etc/tuned/profile_mode", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source", "_original_basename": "tmpm3xpwwsg", "follow": false, "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10000 1726776656.42957: done with _execute_module (ansible.legacy.copy, {'dest': '/etc/tuned/profile_mode', 'mode': '0600', 'src': '/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source', '_original_basename': 'tmpm3xpwwsg', 'follow': False, 'checksum': '3ef9f23deed2e23d3ef2b88b842fb882313e15ce', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10000 1726776656.42967: _low_level_execute_command(): starting 10000 1726776656.42972: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/ > /dev/null 2>&1 && sleep 0' 10000 1726776656.45557: stderr chunk (state=2): >>><<< 10000 1726776656.45566: stdout chunk (state=2): >>><<< 10000 1726776656.45577: _low_level_execute_command() done: rc=0, stdout=, stderr= 10000 1726776656.45582: handler run complete 10000 1726776656.45598: attempt loop complete, returning result 10000 1726776656.45600: _execute() done 10000 1726776656.45602: dumping result to json 10000 1726776656.45605: done dumping result, returning 10000 1726776656.45610: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual [120fa90a-8a95-fee7-0cea-000000000057] 10000 1726776656.45614: sending task result for task 120fa90a-8a95-fee7-0cea-000000000057 10000 1726776656.45648: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000057 10000 1726776656.45650: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "dest": "/etc/tuned/profile_mode", "gid": 0, "group": "root", "md5sum": "cf3f2a865fbea819dadd439586eaee31", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "src": "/root/.ansible/tmp/ansible-tmp-1726776655.8803492-10000-129868350717347/source", "state": "file", "uid": 0 } 9704 1726776656.45841: no more pending results, returning what we have 9704 1726776656.45844: results queue empty 9704 1726776656.45845: checking for any_errors_fatal 9704 1726776656.45850: done checking for any_errors_fatal 9704 1726776656.45850: checking for max_fail_percentage 9704 1726776656.45852: done checking for max_fail_percentage 9704 1726776656.45852: checking to see if all hosts have failed and the running result is not ok 9704 1726776656.45853: done checking to see if all hosts have failed 9704 1726776656.45853: getting the remaining hosts for this loop 9704 1726776656.45854: done getting the remaining hosts for this loop 9704 1726776656.45858: getting the next task for host managed_node3 9704 1726776656.45862: done getting next task for host managed_node3 9704 1726776656.45866: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get current config 9704 1726776656.45868: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776656.45876: getting variables 9704 1726776656.45878: in VariableManager get_vars() 9704 1726776656.45906: Calling all_inventory to load vars for managed_node3 9704 1726776656.45908: Calling groups_inventory to load vars for managed_node3 9704 1726776656.45910: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776656.45916: Calling all_plugins_play to load vars for managed_node3 9704 1726776656.45918: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776656.45919: Calling groups_plugins_play to load vars for managed_node3 9704 1726776656.46025: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776656.46146: done with get_vars() 9704 1726776656.46155: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Get current config] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:107 Thursday 19 September 2024 16:10:56 -0400 (0:00:00.622) 0:00:09.390 **** 9704 1726776656.46211: entering _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776656.46379: worker is 1 (out of 1 available) 9704 1726776656.46392: exiting _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776656.46404: done queuing things up, now waiting for results queue to drain 9704 1726776656.46406: waiting for pending results... 10028 1726776656.46521: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get current config 10028 1726776656.46630: in run() - task 120fa90a-8a95-fee7-0cea-000000000058 10028 1726776656.46646: variable 'ansible_search_path' from source: unknown 10028 1726776656.46650: variable 'ansible_search_path' from source: unknown 10028 1726776656.46678: calling self._execute() 10028 1726776656.46752: variable 'ansible_host' from source: host vars for 'managed_node3' 10028 1726776656.46761: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10028 1726776656.46768: variable 'omit' from source: magic vars 10028 1726776656.46852: variable 'omit' from source: magic vars 10028 1726776656.46883: variable 'omit' from source: magic vars 10028 1726776656.46902: variable '__kernel_settings_profile_filename' from source: role '' all vars 10028 1726776656.47115: variable '__kernel_settings_profile_filename' from source: role '' all vars 10028 1726776656.47174: variable '__kernel_settings_profile_dir' from source: role '' all vars 10028 1726776656.47243: variable '__kernel_settings_profile_parent' from source: set_fact 10028 1726776656.47252: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10028 1726776656.47341: variable 'omit' from source: magic vars 10028 1726776656.47372: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10028 1726776656.47399: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10028 1726776656.47418: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10028 1726776656.47474: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10028 1726776656.47486: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10028 1726776656.47514: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10028 1726776656.47522: variable 'ansible_host' from source: host vars for 'managed_node3' 10028 1726776656.47527: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10028 1726776656.47622: Set connection var ansible_module_compression to ZIP_DEFLATED 10028 1726776656.47630: Set connection var ansible_shell_type to sh 10028 1726776656.47639: Set connection var ansible_timeout to 10 10028 1726776656.47645: Set connection var ansible_pipelining to False 10028 1726776656.47652: Set connection var ansible_shell_executable to /bin/sh 10028 1726776656.47655: Set connection var ansible_connection to ssh 10028 1726776656.47670: variable 'ansible_shell_executable' from source: unknown 10028 1726776656.47673: variable 'ansible_connection' from source: unknown 10028 1726776656.47676: variable 'ansible_module_compression' from source: unknown 10028 1726776656.47679: variable 'ansible_shell_type' from source: unknown 10028 1726776656.47683: variable 'ansible_shell_executable' from source: unknown 10028 1726776656.47686: variable 'ansible_host' from source: host vars for 'managed_node3' 10028 1726776656.47690: variable 'ansible_pipelining' from source: unknown 10028 1726776656.47694: variable 'ansible_timeout' from source: unknown 10028 1726776656.47698: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10028 1726776656.47819: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10028 1726776656.47834: variable 'omit' from source: magic vars 10028 1726776656.47840: starting attempt loop 10028 1726776656.47844: running the handler 10028 1726776656.47855: _low_level_execute_command(): starting 10028 1726776656.47863: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10028 1726776656.50288: stdout chunk (state=2): >>>/root <<< 10028 1726776656.50388: stderr chunk (state=3): >>><<< 10028 1726776656.50394: stdout chunk (state=3): >>><<< 10028 1726776656.50412: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10028 1726776656.50427: _low_level_execute_command(): starting 10028 1726776656.50435: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878 `" && echo ansible-tmp-1726776656.5042202-10028-202561113701878="` echo /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878 `" ) && sleep 0' 10028 1726776656.53238: stdout chunk (state=2): >>>ansible-tmp-1726776656.5042202-10028-202561113701878=/root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878 <<< 10028 1726776656.53354: stderr chunk (state=3): >>><<< 10028 1726776656.53361: stdout chunk (state=3): >>><<< 10028 1726776656.53374: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776656.5042202-10028-202561113701878=/root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878 , stderr= 10028 1726776656.53408: variable 'ansible_module_compression' from source: unknown 10028 1726776656.53441: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.kernel_settings_get_config-ZIP_DEFLATED 10028 1726776656.53471: variable 'ansible_facts' from source: unknown 10028 1726776656.53536: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/AnsiballZ_kernel_settings_get_config.py 10028 1726776656.53864: Sending initial data 10028 1726776656.53871: Sent initial data (174 bytes) 10028 1726776656.56131: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpwmh2xy2c /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/AnsiballZ_kernel_settings_get_config.py <<< 10028 1726776656.57097: stderr chunk (state=3): >>><<< 10028 1726776656.57106: stdout chunk (state=3): >>><<< 10028 1726776656.57131: done transferring module to remote 10028 1726776656.57144: _low_level_execute_command(): starting 10028 1726776656.57150: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/ /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10028 1726776656.59498: stderr chunk (state=2): >>><<< 10028 1726776656.59507: stdout chunk (state=2): >>><<< 10028 1726776656.59523: _low_level_execute_command() done: rc=0, stdout=, stderr= 10028 1726776656.59528: _low_level_execute_command(): starting 10028 1726776656.59534: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10028 1726776656.74978: stdout chunk (state=2): >>> {"changed": false, "data": {}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}} <<< 10028 1726776656.76019: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10028 1726776656.76066: stderr chunk (state=3): >>><<< 10028 1726776656.76078: stdout chunk (state=3): >>><<< 10028 1726776656.76093: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "data": {}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 10028 1726776656.76112: done with _execute_module (fedora.linux_system_roles.kernel_settings_get_config, {'path': '/etc/tuned/kernel_settings/tuned.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'fedora.linux_system_roles.kernel_settings_get_config', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10028 1726776656.76123: _low_level_execute_command(): starting 10028 1726776656.76132: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776656.5042202-10028-202561113701878/ > /dev/null 2>&1 && sleep 0' 10028 1726776656.78617: stderr chunk (state=2): >>><<< 10028 1726776656.78625: stdout chunk (state=2): >>><<< 10028 1726776656.78641: _low_level_execute_command() done: rc=0, stdout=, stderr= 10028 1726776656.78647: handler run complete 10028 1726776656.78660: attempt loop complete, returning result 10028 1726776656.78664: _execute() done 10028 1726776656.78667: dumping result to json 10028 1726776656.78672: done dumping result, returning 10028 1726776656.78679: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get current config [120fa90a-8a95-fee7-0cea-000000000058] 10028 1726776656.78684: sending task result for task 120fa90a-8a95-fee7-0cea-000000000058 10028 1726776656.78712: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000058 10028 1726776656.78716: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "data": {} } 9704 1726776656.78845: no more pending results, returning what we have 9704 1726776656.78848: results queue empty 9704 1726776656.78848: checking for any_errors_fatal 9704 1726776656.78854: done checking for any_errors_fatal 9704 1726776656.78855: checking for max_fail_percentage 9704 1726776656.78856: done checking for max_fail_percentage 9704 1726776656.78857: checking to see if all hosts have failed and the running result is not ok 9704 1726776656.78857: done checking to see if all hosts have failed 9704 1726776656.78858: getting the remaining hosts for this loop 9704 1726776656.78859: done getting the remaining hosts for this loop 9704 1726776656.78862: getting the next task for host managed_node3 9704 1726776656.78867: done getting next task for host managed_node3 9704 1726776656.78870: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings 9704 1726776656.78872: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776656.78880: getting variables 9704 1726776656.78882: in VariableManager get_vars() 9704 1726776656.78913: Calling all_inventory to load vars for managed_node3 9704 1726776656.78916: Calling groups_inventory to load vars for managed_node3 9704 1726776656.78918: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776656.78929: Calling all_plugins_play to load vars for managed_node3 9704 1726776656.78932: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776656.78934: Calling groups_plugins_play to load vars for managed_node3 9704 1726776656.79096: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776656.79215: done with get_vars() 9704 1726776656.79225: done getting variables 9704 1726776656.79319: Loading ActionModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/template.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Apply kernel settings] ******* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:112 Thursday 19 September 2024 16:10:56 -0400 (0:00:00.331) 0:00:09.722 **** 9704 1726776656.79344: entering _queue_task() for managed_node3/template 9704 1726776656.79345: Creating lock for template 9704 1726776656.79518: worker is 1 (out of 1 available) 9704 1726776656.79536: exiting _queue_task() for managed_node3/template 9704 1726776656.79548: done queuing things up, now waiting for results queue to drain 9704 1726776656.79550: waiting for pending results... 10047 1726776656.79657: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings 10047 1726776656.79760: in run() - task 120fa90a-8a95-fee7-0cea-000000000059 10047 1726776656.79777: variable 'ansible_search_path' from source: unknown 10047 1726776656.79781: variable 'ansible_search_path' from source: unknown 10047 1726776656.79808: calling self._execute() 10047 1726776656.79871: variable 'ansible_host' from source: host vars for 'managed_node3' 10047 1726776656.79880: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10047 1726776656.79889: variable 'omit' from source: magic vars 10047 1726776656.79964: variable 'omit' from source: magic vars 10047 1726776656.79997: variable 'omit' from source: magic vars 10047 1726776656.80227: variable '__kernel_settings_profile_src' from source: role '' all vars 10047 1726776656.80238: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10047 1726776656.80291: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10047 1726776656.80310: variable '__kernel_settings_profile_filename' from source: role '' all vars 10047 1726776656.80358: variable '__kernel_settings_profile_filename' from source: role '' all vars 10047 1726776656.80406: variable '__kernel_settings_profile_dir' from source: role '' all vars 10047 1726776656.80469: variable '__kernel_settings_profile_parent' from source: set_fact 10047 1726776656.80477: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10047 1726776656.80499: variable 'omit' from source: magic vars 10047 1726776656.80532: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10047 1726776656.80559: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10047 1726776656.80578: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10047 1726776656.80591: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10047 1726776656.80599: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10047 1726776656.80620: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10047 1726776656.80624: variable 'ansible_host' from source: host vars for 'managed_node3' 10047 1726776656.80626: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10047 1726776656.80692: Set connection var ansible_module_compression to ZIP_DEFLATED 10047 1726776656.80698: Set connection var ansible_shell_type to sh 10047 1726776656.80703: Set connection var ansible_timeout to 10 10047 1726776656.80706: Set connection var ansible_pipelining to False 10047 1726776656.80710: Set connection var ansible_shell_executable to /bin/sh 10047 1726776656.80712: Set connection var ansible_connection to ssh 10047 1726776656.80725: variable 'ansible_shell_executable' from source: unknown 10047 1726776656.80730: variable 'ansible_connection' from source: unknown 10047 1726776656.80733: variable 'ansible_module_compression' from source: unknown 10047 1726776656.80736: variable 'ansible_shell_type' from source: unknown 10047 1726776656.80740: variable 'ansible_shell_executable' from source: unknown 10047 1726776656.80743: variable 'ansible_host' from source: host vars for 'managed_node3' 10047 1726776656.80747: variable 'ansible_pipelining' from source: unknown 10047 1726776656.80750: variable 'ansible_timeout' from source: unknown 10047 1726776656.80754: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10047 1726776656.80844: Loading ActionModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/template.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10047 1726776656.80855: variable 'omit' from source: magic vars 10047 1726776656.80861: starting attempt loop 10047 1726776656.80865: running the handler 10047 1726776656.80875: _low_level_execute_command(): starting 10047 1726776656.80883: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10047 1726776656.83716: stdout chunk (state=2): >>>/root <<< 10047 1726776656.83838: stderr chunk (state=3): >>><<< 10047 1726776656.83845: stdout chunk (state=3): >>><<< 10047 1726776656.83862: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10047 1726776656.83875: _low_level_execute_command(): starting 10047 1726776656.83881: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363 `" && echo ansible-tmp-1726776656.838695-10047-163049773564363="` echo /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363 `" ) && sleep 0' 10047 1726776656.86309: stdout chunk (state=2): >>>ansible-tmp-1726776656.838695-10047-163049773564363=/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363 <<< 10047 1726776656.86433: stderr chunk (state=3): >>><<< 10047 1726776656.86439: stdout chunk (state=3): >>><<< 10047 1726776656.86453: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776656.838695-10047-163049773564363=/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363 , stderr= 10047 1726776656.86467: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks 10047 1726776656.86487: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/kernel_settings.j2 10047 1726776656.86508: variable 'ansible_search_path' from source: unknown 10047 1726776656.87100: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10047 1726776656.88837: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10047 1726776656.88888: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10047 1726776656.88917: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10047 1726776656.88947: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10047 1726776656.88967: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10047 1726776656.89159: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10047 1726776656.89180: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10047 1726776656.89201: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10047 1726776656.89234: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10047 1726776656.89247: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10047 1726776656.89471: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10047 1726776656.89489: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10047 1726776656.89506: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10047 1726776656.89535: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10047 1726776656.89546: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10047 1726776656.89791: variable 'ansible_managed' from source: unknown 10047 1726776656.89798: variable '__sections' from source: task vars 10047 1726776656.89887: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10047 1726776656.89905: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10047 1726776656.89924: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10047 1726776656.89951: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10047 1726776656.89962: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10047 1726776656.90035: variable 'kernel_settings_sysctl' from source: include_vars 10047 1726776656.90047: variable '__kernel_settings_state_empty' from source: role '' all vars 10047 1726776656.90052: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10047 1726776656.90095: variable '__sysctl_old' from source: task vars 10047 1726776656.90142: variable '__sysctl_old' from source: task vars 10047 1726776656.90280: variable 'kernel_settings_purge' from source: role '' defaults 10047 1726776656.90286: variable 'kernel_settings_sysctl' from source: include_vars 10047 1726776656.90295: variable '__kernel_settings_state_empty' from source: role '' all vars 10047 1726776656.90299: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10047 1726776656.90304: variable '__kernel_settings_profile_contents' from source: set_fact 10047 1726776656.90432: variable 'kernel_settings_sysfs' from source: include_vars 10047 1726776656.90443: variable '__kernel_settings_state_empty' from source: role '' all vars 10047 1726776656.90448: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10047 1726776656.90474: variable '__sysfs_old' from source: task vars 10047 1726776656.90518: variable '__sysfs_old' from source: task vars 10047 1726776656.90656: variable 'kernel_settings_purge' from source: role '' defaults 10047 1726776656.90663: variable 'kernel_settings_sysfs' from source: include_vars 10047 1726776656.90671: variable '__kernel_settings_state_empty' from source: role '' all vars 10047 1726776656.90675: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10047 1726776656.90680: variable '__kernel_settings_profile_contents' from source: set_fact 10047 1726776656.90694: variable 'kernel_settings_systemd_cpu_affinity' from source: include_vars 10047 1726776656.90739: variable 'kernel_settings_systemd_cpu_affinity' from source: include_vars 10047 1726776656.90755: variable '__systemd_old' from source: task vars 10047 1726776656.90795: variable '__systemd_old' from source: task vars 10047 1726776656.90924: variable 'kernel_settings_purge' from source: role '' defaults 10047 1726776656.90931: variable 'kernel_settings_systemd_cpu_affinity' from source: include_vars 10047 1726776656.90938: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.90943: variable '__kernel_settings_profile_contents' from source: set_fact 10047 1726776656.90952: variable 'kernel_settings_transparent_hugepages' from source: include_vars 10047 1726776656.90995: variable 'kernel_settings_transparent_hugepages' from source: include_vars 10047 1726776656.91004: variable 'kernel_settings_transparent_hugepages_defrag' from source: include_vars 10047 1726776656.91047: variable 'kernel_settings_transparent_hugepages_defrag' from source: include_vars 10047 1726776656.91061: variable '__trans_huge_old' from source: task vars 10047 1726776656.91102: variable '__trans_huge_old' from source: task vars 10047 1726776656.91247: variable 'kernel_settings_purge' from source: role '' defaults 10047 1726776656.91254: variable 'kernel_settings_transparent_hugepages' from source: include_vars 10047 1726776656.91259: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91264: variable '__kernel_settings_profile_contents' from source: set_fact 10047 1726776656.91273: variable '__trans_defrag_old' from source: task vars 10047 1726776656.91315: variable '__trans_defrag_old' from source: task vars 10047 1726776656.91446: variable 'kernel_settings_purge' from source: role '' defaults 10047 1726776656.91452: variable 'kernel_settings_transparent_hugepages_defrag' from source: include_vars 10047 1726776656.91458: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91463: variable '__kernel_settings_profile_contents' from source: set_fact 10047 1726776656.91475: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91485: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91493: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91500: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91508: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91515: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91524: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91532: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91540: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91546: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91560: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91569: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91576: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91582: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91589: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91595: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91601: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91607: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91614: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.91618: variable '__kernel_settings_state_absent' from source: role '' all vars 10047 1726776656.92047: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10047 1726776656.92089: variable 'ansible_module_compression' from source: unknown 10047 1726776656.92131: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10047 1726776656.92156: variable 'ansible_facts' from source: unknown 10047 1726776656.92223: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_stat.py 10047 1726776656.92311: Sending initial data 10047 1726776656.92317: Sent initial data (151 bytes) 10047 1726776656.94868: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpe3xkyb1q /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_stat.py <<< 10047 1726776656.95854: stderr chunk (state=3): >>><<< 10047 1726776656.95860: stdout chunk (state=3): >>><<< 10047 1726776656.95876: done transferring module to remote 10047 1726776656.95886: _low_level_execute_command(): starting 10047 1726776656.95892: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/ /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_stat.py && sleep 0' 10047 1726776656.98257: stderr chunk (state=2): >>><<< 10047 1726776656.98265: stdout chunk (state=2): >>><<< 10047 1726776656.98280: _low_level_execute_command() done: rc=0, stdout=, stderr= 10047 1726776656.98284: _low_level_execute_command(): starting 10047 1726776656.98289: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_stat.py && sleep 0' 10047 1726776657.13319: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10047 1726776657.14334: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10047 1726776657.14375: stderr chunk (state=3): >>><<< 10047 1726776657.14384: stdout chunk (state=3): >>><<< 10047 1726776657.14398: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10047 1726776657.14420: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/kernel_settings/tuned.conf', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10047 1726776657.14504: Sending initial data 10047 1726776657.14512: Sent initial data (159 bytes) 10047 1726776657.17024: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpfl2ep9kl/kernel_settings.j2 /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source <<< 10047 1726776657.17285: stderr chunk (state=3): >>><<< 10047 1726776657.17292: stdout chunk (state=3): >>><<< 10047 1726776657.17305: _low_level_execute_command(): starting 10047 1726776657.17311: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/ /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source && sleep 0' 10047 1726776657.19619: stderr chunk (state=2): >>><<< 10047 1726776657.19629: stdout chunk (state=2): >>><<< 10047 1726776657.19641: _low_level_execute_command() done: rc=0, stdout=, stderr= 10047 1726776657.19660: variable 'ansible_module_compression' from source: unknown 10047 1726776657.19692: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.copy-ZIP_DEFLATED 10047 1726776657.19713: variable 'ansible_facts' from source: unknown 10047 1726776657.19777: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_copy.py 10047 1726776657.19859: Sending initial data 10047 1726776657.19866: Sent initial data (151 bytes) 10047 1726776657.22285: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpm3lqeoz7 /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_copy.py <<< 10047 1726776657.23265: stderr chunk (state=3): >>><<< 10047 1726776657.23271: stdout chunk (state=3): >>><<< 10047 1726776657.23286: done transferring module to remote 10047 1726776657.23295: _low_level_execute_command(): starting 10047 1726776657.23299: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/ /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_copy.py && sleep 0' 10047 1726776657.25607: stderr chunk (state=2): >>><<< 10047 1726776657.25614: stdout chunk (state=2): >>><<< 10047 1726776657.25628: _low_level_execute_command() done: rc=0, stdout=, stderr= 10047 1726776657.25635: _low_level_execute_command(): starting 10047 1726776657.25640: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/AnsiballZ_copy.py && sleep 0' 10047 1726776657.41976: stdout chunk (state=2): >>> {"dest": "/etc/tuned/kernel_settings/tuned.conf", "src": "/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source", "md5sum": "db02fa7304c082dd9b84c2d712d07ec0", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0644", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 312, "invocation": {"module_args": {"src": "/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source", "dest": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "follow": false, "_original_basename": "kernel_settings.j2", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10047 1726776657.43158: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10047 1726776657.43201: stderr chunk (state=3): >>><<< 10047 1726776657.43208: stdout chunk (state=3): >>><<< 10047 1726776657.43225: _low_level_execute_command() done: rc=0, stdout= {"dest": "/etc/tuned/kernel_settings/tuned.conf", "src": "/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source", "md5sum": "db02fa7304c082dd9b84c2d712d07ec0", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0644", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 312, "invocation": {"module_args": {"src": "/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source", "dest": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "follow": false, "_original_basename": "kernel_settings.j2", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10047 1726776657.43253: done with _execute_module (ansible.legacy.copy, {'src': '/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source', 'dest': '/etc/tuned/kernel_settings/tuned.conf', 'mode': '0644', 'follow': False, '_original_basename': 'kernel_settings.j2', 'checksum': 'ba15904bb90578344fad097ce2f46f9231275eae', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10047 1726776657.43277: _low_level_execute_command(): starting 10047 1726776657.43284: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/ > /dev/null 2>&1 && sleep 0' 10047 1726776657.45641: stderr chunk (state=2): >>><<< 10047 1726776657.45650: stdout chunk (state=2): >>><<< 10047 1726776657.45664: _low_level_execute_command() done: rc=0, stdout=, stderr= 10047 1726776657.45676: handler run complete 10047 1726776657.45695: attempt loop complete, returning result 10047 1726776657.45699: _execute() done 10047 1726776657.45703: dumping result to json 10047 1726776657.45708: done dumping result, returning 10047 1726776657.45717: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings [120fa90a-8a95-fee7-0cea-000000000059] 10047 1726776657.45726: sending task result for task 120fa90a-8a95-fee7-0cea-000000000059 10047 1726776657.45770: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000059 10047 1726776657.45774: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "dest": "/etc/tuned/kernel_settings/tuned.conf", "gid": 0, "group": "root", "md5sum": "db02fa7304c082dd9b84c2d712d07ec0", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 312, "src": "/root/.ansible/tmp/ansible-tmp-1726776656.838695-10047-163049773564363/source", "state": "file", "uid": 0 } 9704 1726776657.45947: no more pending results, returning what we have 9704 1726776657.45949: results queue empty 9704 1726776657.45950: checking for any_errors_fatal 9704 1726776657.45955: done checking for any_errors_fatal 9704 1726776657.45956: checking for max_fail_percentage 9704 1726776657.45957: done checking for max_fail_percentage 9704 1726776657.45958: checking to see if all hosts have failed and the running result is not ok 9704 1726776657.45958: done checking to see if all hosts have failed 9704 1726776657.45959: getting the remaining hosts for this loop 9704 1726776657.45960: done getting the remaining hosts for this loop 9704 1726776657.45963: getting the next task for host managed_node3 9704 1726776657.45968: done getting next task for host managed_node3 9704 1726776657.45970: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes 9704 1726776657.45973: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776657.45983: getting variables 9704 1726776657.45984: in VariableManager get_vars() 9704 1726776657.46014: Calling all_inventory to load vars for managed_node3 9704 1726776657.46017: Calling groups_inventory to load vars for managed_node3 9704 1726776657.46018: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776657.46026: Calling all_plugins_play to load vars for managed_node3 9704 1726776657.46030: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776657.46032: Calling groups_plugins_play to load vars for managed_node3 9704 1726776657.46147: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776657.46266: done with get_vars() 9704 1726776657.46274: done getting variables 9704 1726776657.46313: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:149 Thursday 19 September 2024 16:10:57 -0400 (0:00:00.669) 0:00:10.392 **** 9704 1726776657.46338: entering _queue_task() for managed_node3/service 9704 1726776657.46504: worker is 1 (out of 1 available) 9704 1726776657.46517: exiting _queue_task() for managed_node3/service 9704 1726776657.46530: done queuing things up, now waiting for results queue to drain 9704 1726776657.46532: waiting for pending results... 10079 1726776657.46650: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes 10079 1726776657.46756: in run() - task 120fa90a-8a95-fee7-0cea-00000000005a 10079 1726776657.46775: variable 'ansible_search_path' from source: unknown 10079 1726776657.46780: variable 'ansible_search_path' from source: unknown 10079 1726776657.46813: variable '__kernel_settings_services' from source: include_vars 10079 1726776657.47132: variable '__kernel_settings_services' from source: include_vars 10079 1726776657.47197: variable 'omit' from source: magic vars 10079 1726776657.47288: variable 'ansible_host' from source: host vars for 'managed_node3' 10079 1726776657.47298: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10079 1726776657.47308: variable 'omit' from source: magic vars 10079 1726776657.47564: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10079 1726776657.47799: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10079 1726776657.47846: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10079 1726776657.47876: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10079 1726776657.47910: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10079 1726776657.48009: variable '__kernel_settings_register_profile' from source: set_fact 10079 1726776657.48027: variable '__kernel_settings_register_mode' from source: set_fact 10079 1726776657.48050: Evaluated conditional (__kernel_settings_register_profile is changed or __kernel_settings_register_mode is changed): True 10079 1726776657.48059: variable 'omit' from source: magic vars 10079 1726776657.48096: variable 'omit' from source: magic vars 10079 1726776657.48144: variable 'item' from source: unknown 10079 1726776657.48211: variable 'item' from source: unknown 10079 1726776657.48239: variable 'omit' from source: magic vars 10079 1726776657.48269: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10079 1726776657.48292: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10079 1726776657.48313: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10079 1726776657.48326: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10079 1726776657.48340: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10079 1726776657.48369: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10079 1726776657.48374: variable 'ansible_host' from source: host vars for 'managed_node3' 10079 1726776657.48378: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10079 1726776657.48461: Set connection var ansible_module_compression to ZIP_DEFLATED 10079 1726776657.48468: Set connection var ansible_shell_type to sh 10079 1726776657.48476: Set connection var ansible_timeout to 10 10079 1726776657.48481: Set connection var ansible_pipelining to False 10079 1726776657.48487: Set connection var ansible_shell_executable to /bin/sh 10079 1726776657.48490: Set connection var ansible_connection to ssh 10079 1726776657.48505: variable 'ansible_shell_executable' from source: unknown 10079 1726776657.48508: variable 'ansible_connection' from source: unknown 10079 1726776657.48511: variable 'ansible_module_compression' from source: unknown 10079 1726776657.48513: variable 'ansible_shell_type' from source: unknown 10079 1726776657.48516: variable 'ansible_shell_executable' from source: unknown 10079 1726776657.48519: variable 'ansible_host' from source: host vars for 'managed_node3' 10079 1726776657.48525: variable 'ansible_pipelining' from source: unknown 10079 1726776657.48528: variable 'ansible_timeout' from source: unknown 10079 1726776657.48533: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10079 1726776657.48620: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10079 1726776657.48676: variable 'omit' from source: magic vars 10079 1726776657.48684: starting attempt loop 10079 1726776657.48687: running the handler 10079 1726776657.48769: variable 'ansible_facts' from source: unknown 10079 1726776657.48900: _low_level_execute_command(): starting 10079 1726776657.48909: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10079 1726776657.51300: stdout chunk (state=2): >>>/root <<< 10079 1726776657.51741: stderr chunk (state=3): >>><<< 10079 1726776657.51749: stdout chunk (state=3): >>><<< 10079 1726776657.51767: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10079 1726776657.51782: _low_level_execute_command(): starting 10079 1726776657.51789: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502 `" && echo ansible-tmp-1726776657.5177565-10079-8233410402502="` echo /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502 `" ) && sleep 0' 10079 1726776657.54266: stdout chunk (state=2): >>>ansible-tmp-1726776657.5177565-10079-8233410402502=/root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502 <<< 10079 1726776657.54396: stderr chunk (state=3): >>><<< 10079 1726776657.54403: stdout chunk (state=3): >>><<< 10079 1726776657.54418: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776657.5177565-10079-8233410402502=/root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502 , stderr= 10079 1726776657.54447: variable 'ansible_module_compression' from source: unknown 10079 1726776657.54491: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 10079 1726776657.54545: variable 'ansible_facts' from source: unknown 10079 1726776657.54702: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/AnsiballZ_systemd.py 10079 1726776657.54809: Sending initial data 10079 1726776657.54817: Sent initial data (153 bytes) 10079 1726776657.57320: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp6cklf1bs /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/AnsiballZ_systemd.py <<< 10079 1726776657.59213: stderr chunk (state=3): >>><<< 10079 1726776657.59226: stdout chunk (state=3): >>><<< 10079 1726776657.59250: done transferring module to remote 10079 1726776657.59262: _low_level_execute_command(): starting 10079 1726776657.59269: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/ /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/AnsiballZ_systemd.py && sleep 0' 10079 1726776657.61641: stderr chunk (state=2): >>><<< 10079 1726776657.61650: stdout chunk (state=2): >>><<< 10079 1726776657.61665: _low_level_execute_command() done: rc=0, stdout=, stderr= 10079 1726776657.61670: _low_level_execute_command(): starting 10079 1726776657.61676: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/AnsiballZ_systemd.py && sleep 0' 10079 1726776658.13732: stdout chunk (state=2): >>> {"name": "tuned", "changed": true, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:31 EDT", "WatchdogTimestampMonotonic": "239516555", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "9800", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ExecMainStartTimestampMonotonic": "239379769", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9800", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:31 EDT] ; stop_time=[n/a] ; pid=9800 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15011840", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHi<<< 10079 1726776658.13750: stdout chunk (state=3): >>>gh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:31 EDT", "StateChangeTimestampMonotonic": "239516558", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveExitTimestampMonotonic": "239379946", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveEnterTimestampMonotonic": "239516558", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveExitTimestampMonotonic": "239266126", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveEnterTimestampMonotonic": "239376765", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ConditionTimestampMonotonic": "239377966", "AssertTimestamp": "Thu 2024-09-19 16:10:31 EDT", "AssertTimestampMonotonic": "239377967", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "d22f365eaba74b3a87958cacc5d42cbe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "restarted", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 10079 1726776658.15461: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10079 1726776658.15505: stderr chunk (state=3): >>><<< 10079 1726776658.15514: stdout chunk (state=3): >>><<< 10079 1726776658.15537: _low_level_execute_command() done: rc=0, stdout= {"name": "tuned", "changed": true, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:31 EDT", "WatchdogTimestampMonotonic": "239516555", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "9800", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ExecMainStartTimestampMonotonic": "239379769", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9800", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:31 EDT] ; stop_time=[n/a] ; pid=9800 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15011840", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:31 EDT", "StateChangeTimestampMonotonic": "239516558", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveExitTimestampMonotonic": "239379946", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveEnterTimestampMonotonic": "239516558", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveExitTimestampMonotonic": "239266126", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveEnterTimestampMonotonic": "239376765", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ConditionTimestampMonotonic": "239377966", "AssertTimestamp": "Thu 2024-09-19 16:10:31 EDT", "AssertTimestampMonotonic": "239377967", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "d22f365eaba74b3a87958cacc5d42cbe", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "restarted", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10079 1726776658.15641: done with _execute_module (ansible.legacy.systemd, {'name': 'tuned', 'state': 'restarted', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10079 1726776658.15660: _low_level_execute_command(): starting 10079 1726776658.15666: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776657.5177565-10079-8233410402502/ > /dev/null 2>&1 && sleep 0' 10079 1726776658.18388: stderr chunk (state=2): >>><<< 10079 1726776658.18397: stdout chunk (state=2): >>><<< 10079 1726776658.18413: _low_level_execute_command() done: rc=0, stdout=, stderr= 10079 1726776658.18421: handler run complete 10079 1726776658.18475: attempt loop complete, returning result 10079 1726776658.18495: variable 'item' from source: unknown 10079 1726776658.18575: variable 'item' from source: unknown changed: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": true, "enabled": true, "item": "tuned", "name": "tuned", "state": "started", "status": { "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveEnterTimestampMonotonic": "239516558", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ActiveExitTimestampMonotonic": "239266126", "ActiveState": "active", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Thu 2024-09-19 16:10:31 EDT", "AssertTimestampMonotonic": "239377967", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "com.redhat.tuned", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ConditionTimestampMonotonic": "239377966", "ConfigurationDirectoryMode": "0755", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "ControlGroup": "/system.slice/tuned.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Dynamic System Tuning Daemon", "DevicePolicy": "auto", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "9800", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:31 EDT", "ExecMainStartTimestampMonotonic": "239379769", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:31 EDT] ; stop_time=[n/a] ; pid=9800 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "tuned.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveEnterTimestampMonotonic": "239376765", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:31 EDT", "InactiveExitTimestampMonotonic": "239379946", "InvocationID": "d22f365eaba74b3a87958cacc5d42cbe", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "9800", "MemoryAccounting": "yes", "MemoryCurrent": "15011840", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "tuned.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/tuned/tuned.pid", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2024-09-19 16:10:31 EDT", "StateChangeTimestampMonotonic": "239516558", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Thu 2024-09-19 16:10:31 EDT", "WatchdogTimestampMonotonic": "239516555", "WatchdogUSec": "0" } } 10079 1726776658.18716: dumping result to json 10079 1726776658.18738: done dumping result, returning 10079 1726776658.18749: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes [120fa90a-8a95-fee7-0cea-00000000005a] 10079 1726776658.18755: sending task result for task 120fa90a-8a95-fee7-0cea-00000000005a 10079 1726776658.18865: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000005a 10079 1726776658.18870: WORKER PROCESS EXITING 9704 1726776658.19754: no more pending results, returning what we have 9704 1726776658.19757: results queue empty 9704 1726776658.19757: checking for any_errors_fatal 9704 1726776658.19766: done checking for any_errors_fatal 9704 1726776658.19767: checking for max_fail_percentage 9704 1726776658.19768: done checking for max_fail_percentage 9704 1726776658.19769: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.19769: done checking to see if all hosts have failed 9704 1726776658.19770: getting the remaining hosts for this loop 9704 1726776658.19771: done getting the remaining hosts for this loop 9704 1726776658.19774: getting the next task for host managed_node3 9704 1726776658.19779: done getting next task for host managed_node3 9704 1726776658.19782: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings 9704 1726776658.19785: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.19794: getting variables 9704 1726776658.19795: in VariableManager get_vars() 9704 1726776658.19820: Calling all_inventory to load vars for managed_node3 9704 1726776658.19826: Calling groups_inventory to load vars for managed_node3 9704 1726776658.19828: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.19838: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.19841: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.19844: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.19996: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.20215: done with get_vars() 9704 1726776658.20227: done getting variables 9704 1726776658.20313: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Tuned apply settings] ******** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:157 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.740) 0:00:11.132 **** 9704 1726776658.20346: entering _queue_task() for managed_node3/command 9704 1726776658.20348: Creating lock for command 9704 1726776658.20566: worker is 1 (out of 1 available) 9704 1726776658.20580: exiting _queue_task() for managed_node3/command 9704 1726776658.20593: done queuing things up, now waiting for results queue to drain 9704 1726776658.20595: waiting for pending results... 10102 1726776658.20810: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings 10102 1726776658.20951: in run() - task 120fa90a-8a95-fee7-0cea-00000000005b 10102 1726776658.20968: variable 'ansible_search_path' from source: unknown 10102 1726776658.20973: variable 'ansible_search_path' from source: unknown 10102 1726776658.21007: calling self._execute() 10102 1726776658.21087: variable 'ansible_host' from source: host vars for 'managed_node3' 10102 1726776658.21096: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10102 1726776658.21104: variable 'omit' from source: magic vars 10102 1726776658.21553: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10102 1726776658.21810: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10102 1726776658.21851: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10102 1726776658.21885: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10102 1726776658.21919: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10102 1726776658.22013: variable '__kernel_settings_register_profile' from source: set_fact 10102 1726776658.22036: Evaluated conditional (not __kernel_settings_register_profile is changed): False 10102 1726776658.22041: when evaluation is False, skipping this task 10102 1726776658.22043: _execute() done 10102 1726776658.22046: dumping result to json 10102 1726776658.22048: done dumping result, returning 10102 1726776658.22051: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings [120fa90a-8a95-fee7-0cea-00000000005b] 10102 1726776658.22055: sending task result for task 120fa90a-8a95-fee7-0cea-00000000005b 10102 1726776658.22072: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000005b 10102 1726776658.22075: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_register_profile is changed", "skip_reason": "Conditional result was False" } 9704 1726776658.22342: no more pending results, returning what we have 9704 1726776658.22345: results queue empty 9704 1726776658.22345: checking for any_errors_fatal 9704 1726776658.22358: done checking for any_errors_fatal 9704 1726776658.22359: checking for max_fail_percentage 9704 1726776658.22360: done checking for max_fail_percentage 9704 1726776658.22361: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.22361: done checking to see if all hosts have failed 9704 1726776658.22362: getting the remaining hosts for this loop 9704 1726776658.22363: done getting the remaining hosts for this loop 9704 1726776658.22366: getting the next task for host managed_node3 9704 1726776658.22377: done getting next task for host managed_node3 9704 1726776658.22380: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Verify settings 9704 1726776658.22383: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.22393: getting variables 9704 1726776658.22394: in VariableManager get_vars() 9704 1726776658.22423: Calling all_inventory to load vars for managed_node3 9704 1726776658.22427: Calling groups_inventory to load vars for managed_node3 9704 1726776658.22432: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.22440: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.22442: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.22445: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.22601: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.22753: done with get_vars() 9704 1726776658.22760: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Verify settings] ************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:166 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.024) 0:00:11.156 **** 9704 1726776658.22818: entering _queue_task() for managed_node3/include_tasks 9704 1726776658.22960: worker is 1 (out of 1 available) 9704 1726776658.22973: exiting _queue_task() for managed_node3/include_tasks 9704 1726776658.22983: done queuing things up, now waiting for results queue to drain 9704 1726776658.22985: waiting for pending results... 10111 1726776658.23114: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Verify settings 10111 1726776658.23245: in run() - task 120fa90a-8a95-fee7-0cea-00000000005c 10111 1726776658.23260: variable 'ansible_search_path' from source: unknown 10111 1726776658.23263: variable 'ansible_search_path' from source: unknown 10111 1726776658.23291: calling self._execute() 10111 1726776658.23368: variable 'ansible_host' from source: host vars for 'managed_node3' 10111 1726776658.23377: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10111 1726776658.23386: variable 'omit' from source: magic vars 10111 1726776658.23804: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10111 1726776658.24070: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10111 1726776658.24188: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10111 1726776658.24221: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10111 1726776658.24258: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10111 1726776658.24356: variable '__kernel_settings_register_apply' from source: set_fact 10111 1726776658.24380: Evaluated conditional (__kernel_settings_register_apply is changed): True 10111 1726776658.24389: _execute() done 10111 1726776658.24393: dumping result to json 10111 1726776658.24397: done dumping result, returning 10111 1726776658.24402: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Verify settings [120fa90a-8a95-fee7-0cea-00000000005c] 10111 1726776658.24407: sending task result for task 120fa90a-8a95-fee7-0cea-00000000005c 10111 1726776658.24438: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000005c 10111 1726776658.24442: WORKER PROCESS EXITING 9704 1726776658.24749: no more pending results, returning what we have 9704 1726776658.24752: in VariableManager get_vars() 9704 1726776658.24785: Calling all_inventory to load vars for managed_node3 9704 1726776658.24788: Calling groups_inventory to load vars for managed_node3 9704 1726776658.24790: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.24798: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.24800: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.24803: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.24964: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.25209: done with get_vars() 9704 1726776658.25216: variable 'ansible_search_path' from source: unknown 9704 1726776658.25217: variable 'ansible_search_path' from source: unknown 9704 1726776658.25254: we have included files to process 9704 1726776658.25255: generating all_blocks data 9704 1726776658.25258: done generating all_blocks data 9704 1726776658.25262: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml 9704 1726776658.25263: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml 9704 1726776658.25265: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml for managed_node3 9704 1726776658.25744: done processing included file 9704 1726776658.25746: iterating over new_blocks loaded from include file 9704 1726776658.25747: in VariableManager get_vars() 9704 1726776658.25767: done with get_vars() 9704 1726776658.25768: filtering new block on tags 9704 1726776658.25813: done filtering new block on tags 9704 1726776658.25815: done iterating over new_blocks loaded from include file 9704 1726776658.25815: extending task lists for all hosts with included blocks 9704 1726776658.26460: done extending task lists 9704 1726776658.26462: done processing included files 9704 1726776658.26462: results queue empty 9704 1726776658.26463: checking for any_errors_fatal 9704 1726776658.26465: done checking for any_errors_fatal 9704 1726776658.26466: checking for max_fail_percentage 9704 1726776658.26467: done checking for max_fail_percentage 9704 1726776658.26467: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.26468: done checking to see if all hosts have failed 9704 1726776658.26469: getting the remaining hosts for this loop 9704 1726776658.26470: done getting the remaining hosts for this loop 9704 1726776658.26472: getting the next task for host managed_node3 9704 1726776658.26476: done getting next task for host managed_node3 9704 1726776658.26478: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly 9704 1726776658.26480: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.26489: getting variables 9704 1726776658.26490: in VariableManager get_vars() 9704 1726776658.26502: Calling all_inventory to load vars for managed_node3 9704 1726776658.26504: Calling groups_inventory to load vars for managed_node3 9704 1726776658.26506: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.26510: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.26512: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.26515: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.26680: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.26880: done with get_vars() 9704 1726776658.26889: done getting variables 9704 1726776658.26932: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:2 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.041) 0:00:11.198 **** 9704 1726776658.26963: entering _queue_task() for managed_node3/command 9704 1726776658.27199: worker is 1 (out of 1 available) 9704 1726776658.27212: exiting _queue_task() for managed_node3/command 9704 1726776658.27225: done queuing things up, now waiting for results queue to drain 9704 1726776658.27227: waiting for pending results... 10116 1726776658.27442: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly 10116 1726776658.27586: in run() - task 120fa90a-8a95-fee7-0cea-00000000012b 10116 1726776658.27603: variable 'ansible_search_path' from source: unknown 10116 1726776658.27608: variable 'ansible_search_path' from source: unknown 10116 1726776658.27645: calling self._execute() 10116 1726776658.27727: variable 'ansible_host' from source: host vars for 'managed_node3' 10116 1726776658.27737: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10116 1726776658.27745: variable 'omit' from source: magic vars 10116 1726776658.27840: variable 'omit' from source: magic vars 10116 1726776658.27898: variable 'omit' from source: magic vars 10116 1726776658.27931: variable 'omit' from source: magic vars 10116 1726776658.27971: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10116 1726776658.28005: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10116 1726776658.28033: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10116 1726776658.28051: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10116 1726776658.28063: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10116 1726776658.28092: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10116 1726776658.28098: variable 'ansible_host' from source: host vars for 'managed_node3' 10116 1726776658.28103: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10116 1726776658.28203: Set connection var ansible_module_compression to ZIP_DEFLATED 10116 1726776658.28212: Set connection var ansible_shell_type to sh 10116 1726776658.28220: Set connection var ansible_timeout to 10 10116 1726776658.28230: Set connection var ansible_pipelining to False 10116 1726776658.28239: Set connection var ansible_shell_executable to /bin/sh 10116 1726776658.28242: Set connection var ansible_connection to ssh 10116 1726776658.28262: variable 'ansible_shell_executable' from source: unknown 10116 1726776658.28267: variable 'ansible_connection' from source: unknown 10116 1726776658.28271: variable 'ansible_module_compression' from source: unknown 10116 1726776658.28274: variable 'ansible_shell_type' from source: unknown 10116 1726776658.28277: variable 'ansible_shell_executable' from source: unknown 10116 1726776658.28279: variable 'ansible_host' from source: host vars for 'managed_node3' 10116 1726776658.28283: variable 'ansible_pipelining' from source: unknown 10116 1726776658.28286: variable 'ansible_timeout' from source: unknown 10116 1726776658.28289: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10116 1726776658.28497: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10116 1726776658.28512: variable 'omit' from source: magic vars 10116 1726776658.28521: starting attempt loop 10116 1726776658.28527: running the handler 10116 1726776658.28542: _low_level_execute_command(): starting 10116 1726776658.28551: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10116 1726776658.31606: stdout chunk (state=2): >>>/root <<< 10116 1726776658.31618: stderr chunk (state=2): >>><<< 10116 1726776658.31635: stdout chunk (state=3): >>><<< 10116 1726776658.31652: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10116 1726776658.31667: _low_level_execute_command(): starting 10116 1726776658.31673: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032 `" && echo ansible-tmp-1726776658.3165958-10116-26127972742032="` echo /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032 `" ) && sleep 0' 10116 1726776658.34627: stdout chunk (state=2): >>>ansible-tmp-1726776658.3165958-10116-26127972742032=/root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032 <<< 10116 1726776658.34797: stderr chunk (state=3): >>><<< 10116 1726776658.34805: stdout chunk (state=3): >>><<< 10116 1726776658.34823: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776658.3165958-10116-26127972742032=/root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032 , stderr= 10116 1726776658.34854: variable 'ansible_module_compression' from source: unknown 10116 1726776658.34909: ANSIBALLZ: Using generic lock for ansible.legacy.command 10116 1726776658.34915: ANSIBALLZ: Acquiring lock 10116 1726776658.34919: ANSIBALLZ: Lock acquired: 140007686719920 10116 1726776658.34922: ANSIBALLZ: Creating module 10116 1726776658.48388: ANSIBALLZ: Writing module into payload 10116 1726776658.48498: ANSIBALLZ: Writing module 10116 1726776658.48519: ANSIBALLZ: Renaming module 10116 1726776658.48526: ANSIBALLZ: Done creating module 10116 1726776658.48544: variable 'ansible_facts' from source: unknown 10116 1726776658.48626: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/AnsiballZ_command.py 10116 1726776658.49539: Sending initial data 10116 1726776658.49546: Sent initial data (154 bytes) 10116 1726776658.52734: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp06n80hb_ /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/AnsiballZ_command.py <<< 10116 1726776658.53534: stderr chunk (state=3): >>><<< 10116 1726776658.53544: stdout chunk (state=3): >>><<< 10116 1726776658.53565: done transferring module to remote 10116 1726776658.53577: _low_level_execute_command(): starting 10116 1726776658.53583: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/ /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/AnsiballZ_command.py && sleep 0' 10116 1726776658.56388: stderr chunk (state=2): >>><<< 10116 1726776658.56397: stdout chunk (state=2): >>><<< 10116 1726776658.56411: _low_level_execute_command() done: rc=0, stdout=, stderr= 10116 1726776658.56415: _low_level_execute_command(): starting 10116 1726776658.56420: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/AnsiballZ_command.py && sleep 0' 10116 1726776658.82436: stdout chunk (state=2): >>> {"changed": true, "stdout": "Verification succeeded, current system settings match the preset profile.\nSee TuneD log file ('/var/log/tuned/tuned.log') for details.", "stderr": "", "rc": 0, "cmd": ["tuned-adm", "verify", "-i"], "start": "2024-09-19 16:10:58.716845", "end": "2024-09-19 16:10:58.822500", "delta": "0:00:00.105655", "msg": "", "invocation": {"module_args": {"_raw_params": "tuned-adm verify -i", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 10116 1726776658.83639: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10116 1726776658.83686: stderr chunk (state=3): >>><<< 10116 1726776658.83694: stdout chunk (state=3): >>><<< 10116 1726776658.83714: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "Verification succeeded, current system settings match the preset profile.\nSee TuneD log file ('/var/log/tuned/tuned.log') for details.", "stderr": "", "rc": 0, "cmd": ["tuned-adm", "verify", "-i"], "start": "2024-09-19 16:10:58.716845", "end": "2024-09-19 16:10:58.822500", "delta": "0:00:00.105655", "msg": "", "invocation": {"module_args": {"_raw_params": "tuned-adm verify -i", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10116 1726776658.83766: done with _execute_module (ansible.legacy.command, {'_raw_params': 'tuned-adm verify -i', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10116 1726776658.83777: _low_level_execute_command(): starting 10116 1726776658.83785: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776658.3165958-10116-26127972742032/ > /dev/null 2>&1 && sleep 0' 10116 1726776658.86426: stderr chunk (state=2): >>><<< 10116 1726776658.86437: stdout chunk (state=2): >>><<< 10116 1726776658.86452: _low_level_execute_command() done: rc=0, stdout=, stderr= 10116 1726776658.86461: handler run complete 10116 1726776658.86485: Evaluated conditional (False): False 10116 1726776658.86498: attempt loop complete, returning result 10116 1726776658.86502: _execute() done 10116 1726776658.86506: dumping result to json 10116 1726776658.86511: done dumping result, returning 10116 1726776658.86518: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly [120fa90a-8a95-fee7-0cea-00000000012b] 10116 1726776658.86525: sending task result for task 120fa90a-8a95-fee7-0cea-00000000012b 10116 1726776658.86564: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000012b 10116 1726776658.86568: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "cmd": [ "tuned-adm", "verify", "-i" ], "delta": "0:00:00.105655", "end": "2024-09-19 16:10:58.822500", "rc": 0, "start": "2024-09-19 16:10:58.716845" } STDOUT: Verification succeeded, current system settings match the preset profile. See TuneD log file ('/var/log/tuned/tuned.log') for details. 9704 1726776658.87044: no more pending results, returning what we have 9704 1726776658.87047: results queue empty 9704 1726776658.87048: checking for any_errors_fatal 9704 1726776658.87049: done checking for any_errors_fatal 9704 1726776658.87050: checking for max_fail_percentage 9704 1726776658.87051: done checking for max_fail_percentage 9704 1726776658.87052: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.87052: done checking to see if all hosts have failed 9704 1726776658.87053: getting the remaining hosts for this loop 9704 1726776658.87054: done getting the remaining hosts for this loop 9704 1726776658.87057: getting the next task for host managed_node3 9704 1726776658.87063: done getting next task for host managed_node3 9704 1726776658.87066: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get last verify results from log 9704 1726776658.87070: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.87079: getting variables 9704 1726776658.87081: in VariableManager get_vars() 9704 1726776658.87115: Calling all_inventory to load vars for managed_node3 9704 1726776658.87118: Calling groups_inventory to load vars for managed_node3 9704 1726776658.87120: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.87129: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.87132: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.87135: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.87299: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.87505: done with get_vars() 9704 1726776658.87515: done getting variables 9704 1726776658.87601: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Get last verify results from log] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:12 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.606) 0:00:11.805 **** 9704 1726776658.87634: entering _queue_task() for managed_node3/shell 9704 1726776658.87635: Creating lock for shell 9704 1726776658.87850: worker is 1 (out of 1 available) 9704 1726776658.87862: exiting _queue_task() for managed_node3/shell 9704 1726776658.87872: done queuing things up, now waiting for results queue to drain 9704 1726776658.87875: waiting for pending results... 10156 1726776658.88130: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get last verify results from log 10156 1726776658.88275: in run() - task 120fa90a-8a95-fee7-0cea-00000000012c 10156 1726776658.88292: variable 'ansible_search_path' from source: unknown 10156 1726776658.88296: variable 'ansible_search_path' from source: unknown 10156 1726776658.88326: calling self._execute() 10156 1726776658.88402: variable 'ansible_host' from source: host vars for 'managed_node3' 10156 1726776658.88411: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10156 1726776658.88420: variable 'omit' from source: magic vars 10156 1726776658.88897: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10156 1726776658.89188: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10156 1726776658.89233: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10156 1726776658.89269: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10156 1726776658.89302: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10156 1726776658.89402: variable '__kernel_settings_register_verify_values' from source: set_fact 10156 1726776658.89430: Evaluated conditional (__kernel_settings_register_verify_values is failed): False 10156 1726776658.89436: when evaluation is False, skipping this task 10156 1726776658.89440: _execute() done 10156 1726776658.89443: dumping result to json 10156 1726776658.89447: done dumping result, returning 10156 1726776658.89452: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get last verify results from log [120fa90a-8a95-fee7-0cea-00000000012c] 10156 1726776658.89458: sending task result for task 120fa90a-8a95-fee7-0cea-00000000012c 10156 1726776658.89486: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000012c 10156 1726776658.89490: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify_values is failed", "skip_reason": "Conditional result was False" } 9704 1726776658.90079: no more pending results, returning what we have 9704 1726776658.90083: results queue empty 9704 1726776658.90083: checking for any_errors_fatal 9704 1726776658.90093: done checking for any_errors_fatal 9704 1726776658.90094: checking for max_fail_percentage 9704 1726776658.90095: done checking for max_fail_percentage 9704 1726776658.90096: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.90096: done checking to see if all hosts have failed 9704 1726776658.90097: getting the remaining hosts for this loop 9704 1726776658.90098: done getting the remaining hosts for this loop 9704 1726776658.90101: getting the next task for host managed_node3 9704 1726776658.90107: done getting next task for host managed_node3 9704 1726776658.90110: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors 9704 1726776658.90114: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.90126: getting variables 9704 1726776658.90127: in VariableManager get_vars() 9704 1726776658.90159: Calling all_inventory to load vars for managed_node3 9704 1726776658.90162: Calling groups_inventory to load vars for managed_node3 9704 1726776658.90164: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.90171: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.90174: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.90177: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.90374: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.90569: done with get_vars() 9704 1726776658.90578: done getting variables 9704 1726776658.90631: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:23 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.030) 0:00:11.835 **** 9704 1726776658.90661: entering _queue_task() for managed_node3/fail 9704 1726776658.90879: worker is 1 (out of 1 available) 9704 1726776658.90892: exiting _queue_task() for managed_node3/fail 9704 1726776658.90902: done queuing things up, now waiting for results queue to drain 9704 1726776658.90904: waiting for pending results... 10158 1726776658.91190: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors 10158 1726776658.91307: in run() - task 120fa90a-8a95-fee7-0cea-00000000012d 10158 1726776658.91322: variable 'ansible_search_path' from source: unknown 10158 1726776658.91331: variable 'ansible_search_path' from source: unknown 10158 1726776658.91360: calling self._execute() 10158 1726776658.91416: variable 'ansible_host' from source: host vars for 'managed_node3' 10158 1726776658.91423: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10158 1726776658.91433: variable 'omit' from source: magic vars 10158 1726776658.91780: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10158 1726776658.91954: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10158 1726776658.91996: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10158 1726776658.92030: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10158 1726776658.92085: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10158 1726776658.92185: variable '__kernel_settings_register_verify_values' from source: set_fact 10158 1726776658.92209: Evaluated conditional (__kernel_settings_register_verify_values is failed): False 10158 1726776658.92214: when evaluation is False, skipping this task 10158 1726776658.92217: _execute() done 10158 1726776658.92220: dumping result to json 10158 1726776658.92223: done dumping result, returning 10158 1726776658.92232: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors [120fa90a-8a95-fee7-0cea-00000000012d] 10158 1726776658.92238: sending task result for task 120fa90a-8a95-fee7-0cea-00000000012d 10158 1726776658.92263: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000012d 10158 1726776658.92266: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify_values is failed", "skip_reason": "Conditional result was False" } 9704 1726776658.92865: no more pending results, returning what we have 9704 1726776658.92868: results queue empty 9704 1726776658.92869: checking for any_errors_fatal 9704 1726776658.92874: done checking for any_errors_fatal 9704 1726776658.92874: checking for max_fail_percentage 9704 1726776658.92876: done checking for max_fail_percentage 9704 1726776658.92876: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.92877: done checking to see if all hosts have failed 9704 1726776658.92877: getting the remaining hosts for this loop 9704 1726776658.92878: done getting the remaining hosts for this loop 9704 1726776658.92881: getting the next task for host managed_node3 9704 1726776658.92888: done getting next task for host managed_node3 9704 1726776658.92892: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes 9704 1726776658.92894: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.92906: getting variables 9704 1726776658.92908: in VariableManager get_vars() 9704 1726776658.92939: Calling all_inventory to load vars for managed_node3 9704 1726776658.92942: Calling groups_inventory to load vars for managed_node3 9704 1726776658.92944: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.92952: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.92955: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.92957: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.93119: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.93354: done with get_vars() 9704 1726776658.93364: done getting variables 9704 1726776658.93419: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:177 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.027) 0:00:11.863 **** 9704 1726776658.93451: entering _queue_task() for managed_node3/set_fact 9704 1726776658.93658: worker is 1 (out of 1 available) 9704 1726776658.93671: exiting _queue_task() for managed_node3/set_fact 9704 1726776658.93683: done queuing things up, now waiting for results queue to drain 9704 1726776658.93686: waiting for pending results... 10161 1726776658.93983: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes 10161 1726776658.94091: in run() - task 120fa90a-8a95-fee7-0cea-00000000005d 10161 1726776658.94107: variable 'ansible_search_path' from source: unknown 10161 1726776658.94110: variable 'ansible_search_path' from source: unknown 10161 1726776658.94146: calling self._execute() 10161 1726776658.94222: variable 'ansible_host' from source: host vars for 'managed_node3' 10161 1726776658.94237: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10161 1726776658.94247: variable 'omit' from source: magic vars 10161 1726776658.94344: variable 'omit' from source: magic vars 10161 1726776658.94393: variable 'omit' from source: magic vars 10161 1726776658.94426: variable 'omit' from source: magic vars 10161 1726776658.94467: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10161 1726776658.94502: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10161 1726776658.94526: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10161 1726776658.94546: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10161 1726776658.94559: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10161 1726776658.94589: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10161 1726776658.94595: variable 'ansible_host' from source: host vars for 'managed_node3' 10161 1726776658.94599: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10161 1726776658.94700: Set connection var ansible_module_compression to ZIP_DEFLATED 10161 1726776658.94710: Set connection var ansible_shell_type to sh 10161 1726776658.94718: Set connection var ansible_timeout to 10 10161 1726776658.94727: Set connection var ansible_pipelining to False 10161 1726776658.94738: Set connection var ansible_shell_executable to /bin/sh 10161 1726776658.94742: Set connection var ansible_connection to ssh 10161 1726776658.94762: variable 'ansible_shell_executable' from source: unknown 10161 1726776658.94766: variable 'ansible_connection' from source: unknown 10161 1726776658.94770: variable 'ansible_module_compression' from source: unknown 10161 1726776658.94773: variable 'ansible_shell_type' from source: unknown 10161 1726776658.94776: variable 'ansible_shell_executable' from source: unknown 10161 1726776658.94778: variable 'ansible_host' from source: host vars for 'managed_node3' 10161 1726776658.94781: variable 'ansible_pipelining' from source: unknown 10161 1726776658.94784: variable 'ansible_timeout' from source: unknown 10161 1726776658.94788: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10161 1726776658.94987: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10161 1726776658.94999: variable 'omit' from source: magic vars 10161 1726776658.95006: starting attempt loop 10161 1726776658.95009: running the handler 10161 1726776658.95019: handler run complete 10161 1726776658.95032: attempt loop complete, returning result 10161 1726776658.95037: _execute() done 10161 1726776658.95040: dumping result to json 10161 1726776658.95043: done dumping result, returning 10161 1726776658.95049: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes [120fa90a-8a95-fee7-0cea-00000000005d] 10161 1726776658.95055: sending task result for task 120fa90a-8a95-fee7-0cea-00000000005d 10161 1726776658.95079: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000005d 10161 1726776658.95082: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "kernel_settings_reboot_required": false }, "changed": false } 9704 1726776658.95488: no more pending results, returning what we have 9704 1726776658.95491: results queue empty 9704 1726776658.95492: checking for any_errors_fatal 9704 1726776658.95497: done checking for any_errors_fatal 9704 1726776658.95498: checking for max_fail_percentage 9704 1726776658.95499: done checking for max_fail_percentage 9704 1726776658.95500: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.95501: done checking to see if all hosts have failed 9704 1726776658.95501: getting the remaining hosts for this loop 9704 1726776658.95502: done getting the remaining hosts for this loop 9704 1726776658.95506: getting the next task for host managed_node3 9704 1726776658.95511: done getting next task for host managed_node3 9704 1726776658.95514: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing 9704 1726776658.95517: ^ state is: HOST STATE: block=2, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.95526: getting variables 9704 1726776658.95528: in VariableManager get_vars() 9704 1726776658.95562: Calling all_inventory to load vars for managed_node3 9704 1726776658.95565: Calling groups_inventory to load vars for managed_node3 9704 1726776658.95567: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.95575: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.95577: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.95580: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.95768: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.95942: done with get_vars() 9704 1726776658.95951: done getting variables 9704 1726776658.95989: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:181 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.025) 0:00:11.888 **** 9704 1726776658.96009: entering _queue_task() for managed_node3/set_fact 9704 1726776658.96181: worker is 1 (out of 1 available) 9704 1726776658.96194: exiting _queue_task() for managed_node3/set_fact 9704 1726776658.96205: done queuing things up, now waiting for results queue to drain 9704 1726776658.96207: waiting for pending results... 10172 1726776658.96437: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing 10172 1726776658.96547: in run() - task 120fa90a-8a95-fee7-0cea-00000000005e 10172 1726776658.96562: variable 'ansible_search_path' from source: unknown 10172 1726776658.96566: variable 'ansible_search_path' from source: unknown 10172 1726776658.96594: calling self._execute() 10172 1726776658.96656: variable 'ansible_host' from source: host vars for 'managed_node3' 10172 1726776658.96662: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10172 1726776658.96668: variable 'omit' from source: magic vars 10172 1726776658.96738: variable 'omit' from source: magic vars 10172 1726776658.96781: variable 'omit' from source: magic vars 10172 1726776658.97138: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10172 1726776658.97375: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10172 1726776658.97420: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10172 1726776658.97455: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10172 1726776658.97488: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10172 1726776658.97617: variable '__kernel_settings_register_profile' from source: set_fact 10172 1726776658.97632: variable '__kernel_settings_register_mode' from source: set_fact 10172 1726776658.97641: variable '__kernel_settings_register_apply' from source: set_fact 10172 1726776658.97686: variable 'omit' from source: magic vars 10172 1726776658.97710: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10172 1726776658.97764: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10172 1726776658.97784: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10172 1726776658.97801: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10172 1726776658.97812: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10172 1726776658.97841: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10172 1726776658.97848: variable 'ansible_host' from source: host vars for 'managed_node3' 10172 1726776658.97852: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10172 1726776658.97947: Set connection var ansible_module_compression to ZIP_DEFLATED 10172 1726776658.97955: Set connection var ansible_shell_type to sh 10172 1726776658.97964: Set connection var ansible_timeout to 10 10172 1726776658.97970: Set connection var ansible_pipelining to False 10172 1726776658.97978: Set connection var ansible_shell_executable to /bin/sh 10172 1726776658.97981: Set connection var ansible_connection to ssh 10172 1726776658.97999: variable 'ansible_shell_executable' from source: unknown 10172 1726776658.98004: variable 'ansible_connection' from source: unknown 10172 1726776658.98008: variable 'ansible_module_compression' from source: unknown 10172 1726776658.98011: variable 'ansible_shell_type' from source: unknown 10172 1726776658.98014: variable 'ansible_shell_executable' from source: unknown 10172 1726776658.98017: variable 'ansible_host' from source: host vars for 'managed_node3' 10172 1726776658.98021: variable 'ansible_pipelining' from source: unknown 10172 1726776658.98024: variable 'ansible_timeout' from source: unknown 10172 1726776658.98030: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10172 1726776658.98124: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10172 1726776658.98139: variable 'omit' from source: magic vars 10172 1726776658.98146: starting attempt loop 10172 1726776658.98149: running the handler 10172 1726776658.98160: handler run complete 10172 1726776658.98168: attempt loop complete, returning result 10172 1726776658.98172: _execute() done 10172 1726776658.98175: dumping result to json 10172 1726776658.98178: done dumping result, returning 10172 1726776658.98185: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing [120fa90a-8a95-fee7-0cea-00000000005e] 10172 1726776658.98190: sending task result for task 120fa90a-8a95-fee7-0cea-00000000005e 10172 1726776658.98213: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000005e 10172 1726776658.98217: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_changed": true }, "changed": false } 9704 1726776658.98590: no more pending results, returning what we have 9704 1726776658.98593: results queue empty 9704 1726776658.98594: checking for any_errors_fatal 9704 1726776658.98599: done checking for any_errors_fatal 9704 1726776658.98600: checking for max_fail_percentage 9704 1726776658.98602: done checking for max_fail_percentage 9704 1726776658.98603: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.98604: done checking to see if all hosts have failed 9704 1726776658.98604: getting the remaining hosts for this loop 9704 1726776658.98605: done getting the remaining hosts for this loop 9704 1726776658.98609: getting the next task for host managed_node3 9704 1726776658.98616: done getting next task for host managed_node3 9704 1726776658.98617: ^ task is: TASK: meta (role_complete) 9704 1726776658.98619: ^ state is: HOST STATE: block=2, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.98630: getting variables 9704 1726776658.98631: in VariableManager get_vars() 9704 1726776658.98662: Calling all_inventory to load vars for managed_node3 9704 1726776658.98666: Calling groups_inventory to load vars for managed_node3 9704 1726776658.98668: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.98675: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.98678: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.98680: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.98838: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.99033: done with get_vars() 9704 1726776658.99042: done getting variables 9704 1726776658.99116: done queuing things up, now waiting for results queue to drain 9704 1726776658.99118: results queue empty 9704 1726776658.99119: checking for any_errors_fatal 9704 1726776658.99122: done checking for any_errors_fatal 9704 1726776658.99122: checking for max_fail_percentage 9704 1726776658.99123: done checking for max_fail_percentage 9704 1726776658.99128: checking to see if all hosts have failed and the running result is not ok 9704 1726776658.99130: done checking to see if all hosts have failed 9704 1726776658.99131: getting the remaining hosts for this loop 9704 1726776658.99132: done getting the remaining hosts for this loop 9704 1726776658.99134: getting the next task for host managed_node3 9704 1726776658.99137: done getting next task for host managed_node3 9704 1726776658.99140: ^ task is: TASK: Verify that settings were applied correctly 9704 1726776658.99141: ^ state is: HOST STATE: block=2, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776658.99143: getting variables 9704 1726776658.99144: in VariableManager get_vars() 9704 1726776658.99154: Calling all_inventory to load vars for managed_node3 9704 1726776658.99156: Calling groups_inventory to load vars for managed_node3 9704 1726776658.99158: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776658.99163: Calling all_plugins_play to load vars for managed_node3 9704 1726776658.99165: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776658.99168: Calling groups_plugins_play to load vars for managed_node3 9704 1726776658.99331: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776658.99510: done with get_vars() 9704 1726776658.99518: done getting variables TASK [Verify that settings were applied correctly] ***************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:30 Thursday 19 September 2024 16:10:58 -0400 (0:00:00.035) 0:00:11.924 **** 9704 1726776658.99582: entering _queue_task() for managed_node3/include_tasks 9704 1726776658.99789: worker is 1 (out of 1 available) 9704 1726776658.99800: exiting _queue_task() for managed_node3/include_tasks 9704 1726776658.99811: done queuing things up, now waiting for results queue to drain 9704 1726776658.99813: waiting for pending results... 10174 1726776659.00009: running TaskExecutor() for managed_node3/TASK: Verify that settings were applied correctly 10174 1726776659.00121: in run() - task 120fa90a-8a95-fee7-0cea-000000000009 10174 1726776659.00140: variable 'ansible_search_path' from source: unknown 10174 1726776659.00173: calling self._execute() 10174 1726776659.00251: variable 'ansible_host' from source: host vars for 'managed_node3' 10174 1726776659.00260: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10174 1726776659.00268: variable 'omit' from source: magic vars 10174 1726776659.00354: _execute() done 10174 1726776659.00361: dumping result to json 10174 1726776659.00365: done dumping result, returning 10174 1726776659.00371: done running TaskExecutor() for managed_node3/TASK: Verify that settings were applied correctly [120fa90a-8a95-fee7-0cea-000000000009] 10174 1726776659.00378: sending task result for task 120fa90a-8a95-fee7-0cea-000000000009 10174 1726776659.00404: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000009 10174 1726776659.00408: WORKER PROCESS EXITING 9704 1726776659.00698: no more pending results, returning what we have 9704 1726776659.00702: in VariableManager get_vars() 9704 1726776659.00736: Calling all_inventory to load vars for managed_node3 9704 1726776659.00739: Calling groups_inventory to load vars for managed_node3 9704 1726776659.00741: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.00749: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.00752: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.00754: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.00918: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.01105: done with get_vars() 9704 1726776659.01112: variable 'ansible_search_path' from source: unknown 9704 1726776659.01125: we have included files to process 9704 1726776659.01126: generating all_blocks data 9704 1726776659.01127: done generating all_blocks data 9704 1726776659.01134: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml 9704 1726776659.01135: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml 9704 1726776659.01138: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml 9704 1726776659.01898: in VariableManager get_vars() 9704 1726776659.01916: done with get_vars() included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml for managed_node3 9704 1726776659.02018: done processing included file 9704 1726776659.02019: iterating over new_blocks loaded from include file 9704 1726776659.02020: in VariableManager get_vars() 9704 1726776659.02033: done with get_vars() 9704 1726776659.02034: filtering new block on tags 9704 1726776659.02081: done filtering new block on tags 9704 1726776659.02083: done iterating over new_blocks loaded from include file 9704 1726776659.02083: extending task lists for all hosts with included blocks 9704 1726776659.03080: done extending task lists 9704 1726776659.03081: done processing included files 9704 1726776659.03082: results queue empty 9704 1726776659.03082: checking for any_errors_fatal 9704 1726776659.03084: done checking for any_errors_fatal 9704 1726776659.03084: checking for max_fail_percentage 9704 1726776659.03085: done checking for max_fail_percentage 9704 1726776659.03086: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.03087: done checking to see if all hosts have failed 9704 1726776659.03087: getting the remaining hosts for this loop 9704 1726776659.03088: done getting the remaining hosts for this loop 9704 1726776659.03090: getting the next task for host managed_node3 9704 1726776659.03094: done getting next task for host managed_node3 9704 1726776659.03096: ^ task is: TASK: Set version specific variables 9704 1726776659.03098: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.03100: getting variables 9704 1726776659.03101: in VariableManager get_vars() 9704 1726776659.03111: Calling all_inventory to load vars for managed_node3 9704 1726776659.03113: Calling groups_inventory to load vars for managed_node3 9704 1726776659.03115: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.03120: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.03122: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.03130: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.03284: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.03476: done with get_vars() 9704 1726776659.03485: done getting variables 9704 1726776659.03521: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set version specific variables] ****************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:2 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.039) 0:00:11.964 **** 9704 1726776659.03553: entering _queue_task() for managed_node3/include_vars 9704 1726776659.03812: worker is 1 (out of 1 available) 9704 1726776659.03828: exiting _queue_task() for managed_node3/include_vars 9704 1726776659.03841: done queuing things up, now waiting for results queue to drain 9704 1726776659.03843: waiting for pending results... 10175 1726776659.04144: running TaskExecutor() for managed_node3/TASK: Set version specific variables 10175 1726776659.04271: in run() - task 120fa90a-8a95-fee7-0cea-000000000187 10175 1726776659.04289: variable 'ansible_search_path' from source: unknown 10175 1726776659.04294: variable 'ansible_search_path' from source: unknown 10175 1726776659.04330: calling self._execute() 10175 1726776659.04409: variable 'ansible_host' from source: host vars for 'managed_node3' 10175 1726776659.04419: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10175 1726776659.04432: variable 'omit' from source: magic vars 10175 1726776659.04530: variable 'omit' from source: magic vars 10175 1726776659.04573: variable 'omit' from source: magic vars 10175 1726776659.04923: variable 'ffparams' from source: task vars 10175 1726776659.05041: variable 'ansible_distribution' from source: facts 10175 1726776659.05050: variable 'ansible_distribution_major_version' from source: facts 10175 1726776659.05144: variable 'ansible_distribution' from source: facts 10175 1726776659.05297: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 10175 1726776659.05770: Loaded config def from plugin (lookup/first_found) 10175 1726776659.05778: Loading LookupModule 'first_found' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/first_found.py 10175 1726776659.05811: variable 'ansible_search_path' from source: unknown 10175 1726776659.05820: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings 10175 1726776659.05863: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/files/vars/tests_CentOS_8.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/vars/tests_CentOS_8.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/files/vars/tests_CentOS_8.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_CentOS_8.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/files/vars/tests_CentOS_8.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_CentOS_8.yml 10175 1726776659.05886: variable 'ansible_search_path' from source: unknown 10175 1726776659.05891: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings 10175 1726776659.05923: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/files/vars/tests_CentOS.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/vars/tests_CentOS.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/files/vars/tests_CentOS.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_CentOS.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/files/vars/tests_CentOS.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_CentOS.yml 10175 1726776659.05942: variable 'ansible_search_path' from source: unknown 10175 1726776659.05947: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings 10175 1726776659.05976: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/files/vars/tests_default.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/vars/tests_default.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/files/vars/tests_default.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_default.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/files/vars/tests_default.yml /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_default.yml 10175 1726776659.05996: variable 'omit' from source: magic vars 10175 1726776659.06018: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10175 1726776659.06046: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10175 1726776659.06065: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10175 1726776659.06082: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10175 1726776659.06091: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10175 1726776659.06116: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10175 1726776659.06121: variable 'ansible_host' from source: host vars for 'managed_node3' 10175 1726776659.06128: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10175 1726776659.06234: Set connection var ansible_module_compression to ZIP_DEFLATED 10175 1726776659.06245: Set connection var ansible_shell_type to sh 10175 1726776659.06253: Set connection var ansible_timeout to 10 10175 1726776659.06259: Set connection var ansible_pipelining to False 10175 1726776659.06266: Set connection var ansible_shell_executable to /bin/sh 10175 1726776659.06269: Set connection var ansible_connection to ssh 10175 1726776659.06289: variable 'ansible_shell_executable' from source: unknown 10175 1726776659.06294: variable 'ansible_connection' from source: unknown 10175 1726776659.06297: variable 'ansible_module_compression' from source: unknown 10175 1726776659.06300: variable 'ansible_shell_type' from source: unknown 10175 1726776659.06302: variable 'ansible_shell_executable' from source: unknown 10175 1726776659.06305: variable 'ansible_host' from source: host vars for 'managed_node3' 10175 1726776659.06308: variable 'ansible_pipelining' from source: unknown 10175 1726776659.06312: variable 'ansible_timeout' from source: unknown 10175 1726776659.06315: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10175 1726776659.06409: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10175 1726776659.06420: variable 'omit' from source: magic vars 10175 1726776659.06430: starting attempt loop 10175 1726776659.06433: running the handler 10175 1726776659.06483: handler run complete 10175 1726776659.06494: attempt loop complete, returning result 10175 1726776659.06498: _execute() done 10175 1726776659.06501: dumping result to json 10175 1726776659.06505: done dumping result, returning 10175 1726776659.06510: done running TaskExecutor() for managed_node3/TASK: Set version specific variables [120fa90a-8a95-fee7-0cea-000000000187] 10175 1726776659.06515: sending task result for task 120fa90a-8a95-fee7-0cea-000000000187 10175 1726776659.06548: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000187 10175 1726776659.06552: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_test_python_cmd": "/usr/libexec/platform-python", "__kernel_settings_test_python_pkgs": [ "python3-configobj" ] }, "ansible_included_var_files": [ "/tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/vars/tests_default.yml" ], "changed": false } 9704 1726776659.07253: no more pending results, returning what we have 9704 1726776659.07256: results queue empty 9704 1726776659.07257: checking for any_errors_fatal 9704 1726776659.07258: done checking for any_errors_fatal 9704 1726776659.07259: checking for max_fail_percentage 9704 1726776659.07260: done checking for max_fail_percentage 9704 1726776659.07261: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.07264: done checking to see if all hosts have failed 9704 1726776659.07264: getting the remaining hosts for this loop 9704 1726776659.07265: done getting the remaining hosts for this loop 9704 1726776659.07268: getting the next task for host managed_node3 9704 1726776659.07274: done getting next task for host managed_node3 9704 1726776659.07276: ^ task is: TASK: Reset settings success flag 9704 1726776659.07278: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.07281: getting variables 9704 1726776659.07282: in VariableManager get_vars() 9704 1726776659.07310: Calling all_inventory to load vars for managed_node3 9704 1726776659.07313: Calling groups_inventory to load vars for managed_node3 9704 1726776659.07314: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.07323: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.07329: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.07333: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.07493: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.07685: done with get_vars() 9704 1726776659.07694: done getting variables 9704 1726776659.07752: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Reset settings success flag] ********************************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:13 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.042) 0:00:12.006 **** 9704 1726776659.07778: entering _queue_task() for managed_node3/set_fact 9704 1726776659.07988: worker is 1 (out of 1 available) 9704 1726776659.08002: exiting _queue_task() for managed_node3/set_fact 9704 1726776659.08012: done queuing things up, now waiting for results queue to drain 9704 1726776659.08014: waiting for pending results... 10176 1726776659.08226: running TaskExecutor() for managed_node3/TASK: Reset settings success flag 10176 1726776659.08347: in run() - task 120fa90a-8a95-fee7-0cea-000000000188 10176 1726776659.08364: variable 'ansible_search_path' from source: unknown 10176 1726776659.08369: variable 'ansible_search_path' from source: unknown 10176 1726776659.08397: calling self._execute() 10176 1726776659.08471: variable 'ansible_host' from source: host vars for 'managed_node3' 10176 1726776659.08481: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10176 1726776659.08490: variable 'omit' from source: magic vars 10176 1726776659.08588: variable 'omit' from source: magic vars 10176 1726776659.08627: variable 'omit' from source: magic vars 10176 1726776659.08659: variable 'omit' from source: magic vars 10176 1726776659.08695: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10176 1726776659.08731: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10176 1726776659.08753: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10176 1726776659.09065: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10176 1726776659.09080: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10176 1726776659.09106: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10176 1726776659.09112: variable 'ansible_host' from source: host vars for 'managed_node3' 10176 1726776659.09116: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10176 1726776659.09276: Set connection var ansible_module_compression to ZIP_DEFLATED 10176 1726776659.09285: Set connection var ansible_shell_type to sh 10176 1726776659.09294: Set connection var ansible_timeout to 10 10176 1726776659.09300: Set connection var ansible_pipelining to False 10176 1726776659.09308: Set connection var ansible_shell_executable to /bin/sh 10176 1726776659.09311: Set connection var ansible_connection to ssh 10176 1726776659.09334: variable 'ansible_shell_executable' from source: unknown 10176 1726776659.09339: variable 'ansible_connection' from source: unknown 10176 1726776659.09343: variable 'ansible_module_compression' from source: unknown 10176 1726776659.09347: variable 'ansible_shell_type' from source: unknown 10176 1726776659.09349: variable 'ansible_shell_executable' from source: unknown 10176 1726776659.09353: variable 'ansible_host' from source: host vars for 'managed_node3' 10176 1726776659.09357: variable 'ansible_pipelining' from source: unknown 10176 1726776659.09360: variable 'ansible_timeout' from source: unknown 10176 1726776659.09364: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10176 1726776659.09479: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10176 1726776659.09491: variable 'omit' from source: magic vars 10176 1726776659.09496: starting attempt loop 10176 1726776659.09500: running the handler 10176 1726776659.09509: handler run complete 10176 1726776659.09518: attempt loop complete, returning result 10176 1726776659.09521: _execute() done 10176 1726776659.09527: dumping result to json 10176 1726776659.09532: done dumping result, returning 10176 1726776659.09538: done running TaskExecutor() for managed_node3/TASK: Reset settings success flag [120fa90a-8a95-fee7-0cea-000000000188] 10176 1726776659.09544: sending task result for task 120fa90a-8a95-fee7-0cea-000000000188 10176 1726776659.09566: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000188 10176 1726776659.09569: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_success": true }, "changed": false } 9704 1726776659.09971: no more pending results, returning what we have 9704 1726776659.09975: results queue empty 9704 1726776659.09976: checking for any_errors_fatal 9704 1726776659.09980: done checking for any_errors_fatal 9704 1726776659.09981: checking for max_fail_percentage 9704 1726776659.09982: done checking for max_fail_percentage 9704 1726776659.09983: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.09984: done checking to see if all hosts have failed 9704 1726776659.09984: getting the remaining hosts for this loop 9704 1726776659.09986: done getting the remaining hosts for this loop 9704 1726776659.09989: getting the next task for host managed_node3 9704 1726776659.09994: done getting next task for host managed_node3 9704 1726776659.09995: ^ task is: TASK: Check tuned-adm verify 9704 1726776659.09998: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.10000: getting variables 9704 1726776659.10002: in VariableManager get_vars() 9704 1726776659.10033: Calling all_inventory to load vars for managed_node3 9704 1726776659.10036: Calling groups_inventory to load vars for managed_node3 9704 1726776659.10038: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.10045: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.10048: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.10051: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.10446: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.10639: done with get_vars() 9704 1726776659.10647: done getting variables 9704 1726776659.10696: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check tuned-adm verify] ************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:17 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.029) 0:00:12.036 **** 9704 1726776659.10721: entering _queue_task() for managed_node3/command 9704 1726776659.10922: worker is 1 (out of 1 available) 9704 1726776659.10940: exiting _queue_task() for managed_node3/command 9704 1726776659.10950: done queuing things up, now waiting for results queue to drain 9704 1726776659.10952: waiting for pending results... 10177 1726776659.11233: running TaskExecutor() for managed_node3/TASK: Check tuned-adm verify 10177 1726776659.11352: in run() - task 120fa90a-8a95-fee7-0cea-000000000189 10177 1726776659.11369: variable 'ansible_search_path' from source: unknown 10177 1726776659.11373: variable 'ansible_search_path' from source: unknown 10177 1726776659.11404: calling self._execute() 10177 1726776659.11480: variable 'ansible_host' from source: host vars for 'managed_node3' 10177 1726776659.11490: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10177 1726776659.11498: variable 'omit' from source: magic vars 10177 1726776659.11907: variable '__kernel_settings_test_verify' from source: include params 10177 1726776659.11919: Evaluated conditional (__kernel_settings_test_verify): False 10177 1726776659.11923: when evaluation is False, skipping this task 10177 1726776659.11930: _execute() done 10177 1726776659.11935: dumping result to json 10177 1726776659.11938: done dumping result, returning 10177 1726776659.11942: done running TaskExecutor() for managed_node3/TASK: Check tuned-adm verify [120fa90a-8a95-fee7-0cea-000000000189] 10177 1726776659.11949: sending task result for task 120fa90a-8a95-fee7-0cea-000000000189 10177 1726776659.11976: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000189 10177 1726776659.11980: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_test_verify", "skip_reason": "Conditional result was False" } 9704 1726776659.12318: no more pending results, returning what we have 9704 1726776659.12321: results queue empty 9704 1726776659.12322: checking for any_errors_fatal 9704 1726776659.12329: done checking for any_errors_fatal 9704 1726776659.12330: checking for max_fail_percentage 9704 1726776659.12332: done checking for max_fail_percentage 9704 1726776659.12333: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.12333: done checking to see if all hosts have failed 9704 1726776659.12334: getting the remaining hosts for this loop 9704 1726776659.12335: done getting the remaining hosts for this loop 9704 1726776659.12338: getting the next task for host managed_node3 9704 1726776659.12342: done getting next task for host managed_node3 9704 1726776659.12344: ^ task is: TASK: Check for verify errors 9704 1726776659.12347: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.12349: getting variables 9704 1726776659.12350: in VariableManager get_vars() 9704 1726776659.12378: Calling all_inventory to load vars for managed_node3 9704 1726776659.12381: Calling groups_inventory to load vars for managed_node3 9704 1726776659.12383: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.12392: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.12395: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.12398: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.12564: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.12770: done with get_vars() 9704 1726776659.12778: done getting variables 9704 1726776659.12832: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check for verify errors] ************************************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:24 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.021) 0:00:12.057 **** 9704 1726776659.12859: entering _queue_task() for managed_node3/command 9704 1726776659.13040: worker is 1 (out of 1 available) 9704 1726776659.13051: exiting _queue_task() for managed_node3/command 9704 1726776659.13062: done queuing things up, now waiting for results queue to drain 9704 1726776659.13064: waiting for pending results... 10178 1726776659.13262: running TaskExecutor() for managed_node3/TASK: Check for verify errors 10178 1726776659.13375: in run() - task 120fa90a-8a95-fee7-0cea-00000000018a 10178 1726776659.13391: variable 'ansible_search_path' from source: unknown 10178 1726776659.13395: variable 'ansible_search_path' from source: unknown 10178 1726776659.13427: calling self._execute() 10178 1726776659.13498: variable 'ansible_host' from source: host vars for 'managed_node3' 10178 1726776659.13509: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10178 1726776659.13518: variable 'omit' from source: magic vars 10178 1726776659.13998: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10178 1726776659.14274: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10178 1726776659.14315: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10178 1726776659.14351: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10178 1726776659.14384: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10178 1726776659.14462: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10178 1726776659.14487: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10178 1726776659.14511: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10178 1726776659.14542: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10178 1726776659.14641: variable '__kernel_settings_register_verify' from source: set_fact 10178 1726776659.14655: Evaluated conditional (__kernel_settings_register_verify is defined): True 10178 1726776659.14773: variable '__kernel_settings_register_verify' from source: set_fact 10178 1726776659.14784: Evaluated conditional (__kernel_settings_register_verify is failed): False 10178 1726776659.14789: when evaluation is False, skipping this task 10178 1726776659.14792: _execute() done 10178 1726776659.14795: dumping result to json 10178 1726776659.14798: done dumping result, returning 10178 1726776659.14803: done running TaskExecutor() for managed_node3/TASK: Check for verify errors [120fa90a-8a95-fee7-0cea-00000000018a] 10178 1726776659.14808: sending task result for task 120fa90a-8a95-fee7-0cea-00000000018a 10178 1726776659.14837: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000018a 10178 1726776659.14842: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify is failed", "skip_reason": "Conditional result was False" } 9704 1726776659.15146: no more pending results, returning what we have 9704 1726776659.15149: results queue empty 9704 1726776659.15150: checking for any_errors_fatal 9704 1726776659.15153: done checking for any_errors_fatal 9704 1726776659.15154: checking for max_fail_percentage 9704 1726776659.15155: done checking for max_fail_percentage 9704 1726776659.15156: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.15157: done checking to see if all hosts have failed 9704 1726776659.15157: getting the remaining hosts for this loop 9704 1726776659.15158: done getting the remaining hosts for this loop 9704 1726776659.15161: getting the next task for host managed_node3 9704 1726776659.15166: done getting next task for host managed_node3 9704 1726776659.15168: ^ task is: TASK: Check /proc/cmdline 9704 1726776659.15170: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.15173: getting variables 9704 1726776659.15174: in VariableManager get_vars() 9704 1726776659.15201: Calling all_inventory to load vars for managed_node3 9704 1726776659.15203: Calling groups_inventory to load vars for managed_node3 9704 1726776659.15205: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.15215: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.15218: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.15220: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.15421: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.15612: done with get_vars() 9704 1726776659.15620: done getting variables 9704 1726776659.15672: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check /proc/cmdline] ***************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:31 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.028) 0:00:12.085 **** 9704 1726776659.15697: entering _queue_task() for managed_node3/command 9704 1726776659.15874: worker is 1 (out of 1 available) 9704 1726776659.15887: exiting _queue_task() for managed_node3/command 9704 1726776659.15898: done queuing things up, now waiting for results queue to drain 9704 1726776659.15900: waiting for pending results... 10179 1726776659.16098: running TaskExecutor() for managed_node3/TASK: Check /proc/cmdline 10179 1726776659.16214: in run() - task 120fa90a-8a95-fee7-0cea-00000000018b 10179 1726776659.16234: variable 'ansible_search_path' from source: unknown 10179 1726776659.16239: variable 'ansible_search_path' from source: unknown 10179 1726776659.16269: calling self._execute() 10179 1726776659.16342: variable 'ansible_host' from source: host vars for 'managed_node3' 10179 1726776659.16351: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10179 1726776659.16359: variable 'omit' from source: magic vars 10179 1726776659.16769: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10179 1726776659.17054: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10179 1726776659.17095: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10179 1726776659.17132: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10179 1726776659.17167: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10179 1726776659.17269: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10179 1726776659.17293: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10179 1726776659.17317: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10179 1726776659.17348: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10179 1726776659.17447: variable '__kernel_settings_register_verify' from source: set_fact 10179 1726776659.17461: Evaluated conditional (__kernel_settings_register_verify is defined): True 10179 1726776659.17579: variable '__kernel_settings_register_verify' from source: set_fact 10179 1726776659.17589: Evaluated conditional (__kernel_settings_register_verify is failed): False 10179 1726776659.17594: when evaluation is False, skipping this task 10179 1726776659.17597: _execute() done 10179 1726776659.17600: dumping result to json 10179 1726776659.17604: done dumping result, returning 10179 1726776659.17609: done running TaskExecutor() for managed_node3/TASK: Check /proc/cmdline [120fa90a-8a95-fee7-0cea-00000000018b] 10179 1726776659.17614: sending task result for task 120fa90a-8a95-fee7-0cea-00000000018b 10179 1726776659.17644: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000018b 10179 1726776659.17648: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify is failed", "skip_reason": "Conditional result was False" } 9704 1726776659.17969: no more pending results, returning what we have 9704 1726776659.17972: results queue empty 9704 1726776659.17973: checking for any_errors_fatal 9704 1726776659.17976: done checking for any_errors_fatal 9704 1726776659.17976: checking for max_fail_percentage 9704 1726776659.17978: done checking for max_fail_percentage 9704 1726776659.17979: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.17979: done checking to see if all hosts have failed 9704 1726776659.17980: getting the remaining hosts for this loop 9704 1726776659.17981: done getting the remaining hosts for this loop 9704 1726776659.17984: getting the next task for host managed_node3 9704 1726776659.17988: done getting next task for host managed_node3 9704 1726776659.17990: ^ task is: TASK: Set error flag based on verify 9704 1726776659.17992: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.17995: getting variables 9704 1726776659.17997: in VariableManager get_vars() 9704 1726776659.18026: Calling all_inventory to load vars for managed_node3 9704 1726776659.18031: Calling groups_inventory to load vars for managed_node3 9704 1726776659.18033: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.18041: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.18043: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.18046: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.18205: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.18407: done with get_vars() 9704 1726776659.18417: done getting variables 9704 1726776659.18470: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set error flag based on verify] ****************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:38 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.027) 0:00:12.113 **** 9704 1726776659.18495: entering _queue_task() for managed_node3/set_fact 9704 1726776659.18680: worker is 1 (out of 1 available) 9704 1726776659.18693: exiting _queue_task() for managed_node3/set_fact 9704 1726776659.18703: done queuing things up, now waiting for results queue to drain 9704 1726776659.18705: waiting for pending results... 10180 1726776659.18903: running TaskExecutor() for managed_node3/TASK: Set error flag based on verify 10180 1726776659.19022: in run() - task 120fa90a-8a95-fee7-0cea-00000000018c 10180 1726776659.19042: variable 'ansible_search_path' from source: unknown 10180 1726776659.19047: variable 'ansible_search_path' from source: unknown 10180 1726776659.19076: calling self._execute() 10180 1726776659.19152: variable 'ansible_host' from source: host vars for 'managed_node3' 10180 1726776659.19162: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10180 1726776659.19171: variable 'omit' from source: magic vars 10180 1726776659.19646: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10180 1726776659.19909: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10180 1726776659.19956: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10180 1726776659.19985: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10180 1726776659.20015: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10180 1726776659.20089: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10180 1726776659.20112: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10180 1726776659.20141: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10180 1726776659.20166: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10180 1726776659.20265: variable '__kernel_settings_register_verify' from source: set_fact 10180 1726776659.20281: Evaluated conditional (__kernel_settings_register_verify is defined): True 10180 1726776659.20404: variable '__kernel_settings_register_verify' from source: set_fact 10180 1726776659.20415: Evaluated conditional (__kernel_settings_register_verify is failed): False 10180 1726776659.20420: when evaluation is False, skipping this task 10180 1726776659.20423: _execute() done 10180 1726776659.20431: dumping result to json 10180 1726776659.20434: done dumping result, returning 10180 1726776659.20439: done running TaskExecutor() for managed_node3/TASK: Set error flag based on verify [120fa90a-8a95-fee7-0cea-00000000018c] 10180 1726776659.20444: sending task result for task 120fa90a-8a95-fee7-0cea-00000000018c 10180 1726776659.20469: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000018c 10180 1726776659.20472: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify is failed", "skip_reason": "Conditional result was False" } 9704 1726776659.20802: no more pending results, returning what we have 9704 1726776659.20805: results queue empty 9704 1726776659.20806: checking for any_errors_fatal 9704 1726776659.20810: done checking for any_errors_fatal 9704 1726776659.20810: checking for max_fail_percentage 9704 1726776659.20812: done checking for max_fail_percentage 9704 1726776659.20813: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.20813: done checking to see if all hosts have failed 9704 1726776659.20814: getting the remaining hosts for this loop 9704 1726776659.20815: done getting the remaining hosts for this loop 9704 1726776659.20819: getting the next task for host managed_node3 9704 1726776659.20827: done getting next task for host managed_node3 9704 1726776659.20831: ^ task is: TASK: Check config files 9704 1726776659.20834: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.20837: getting variables 9704 1726776659.20838: in VariableManager get_vars() 9704 1726776659.20867: Calling all_inventory to load vars for managed_node3 9704 1726776659.20870: Calling groups_inventory to load vars for managed_node3 9704 1726776659.20872: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.20882: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.20885: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.20887: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.21096: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.21290: done with get_vars() 9704 1726776659.21299: done getting variables TASK [Check config files] ****************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:45 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.028) 0:00:12.142 **** 9704 1726776659.21385: entering _queue_task() for managed_node3/include_tasks 9704 1726776659.21577: worker is 1 (out of 1 available) 9704 1726776659.21589: exiting _queue_task() for managed_node3/include_tasks 9704 1726776659.21600: done queuing things up, now waiting for results queue to drain 9704 1726776659.21601: waiting for pending results... 10181 1726776659.21809: running TaskExecutor() for managed_node3/TASK: Check config files 10181 1726776659.21917: in run() - task 120fa90a-8a95-fee7-0cea-00000000018d 10181 1726776659.21932: variable 'ansible_search_path' from source: unknown 10181 1726776659.21937: variable 'ansible_search_path' from source: unknown 10181 1726776659.21964: calling self._execute() 10181 1726776659.22023: variable 'ansible_host' from source: host vars for 'managed_node3' 10181 1726776659.22033: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10181 1726776659.22042: variable 'omit' from source: magic vars 10181 1726776659.22345: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10181 1726776659.22514: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10181 1726776659.22555: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10181 1726776659.22590: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10181 1726776659.22622: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10181 1726776659.22727: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10181 1726776659.22857: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10181 1726776659.22886: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10181 1726776659.22913: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10181 1726776659.23028: variable '__kernel_settings_profile_file' from source: include_vars 10181 1726776659.23179: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10181 1726776659.25186: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10181 1726776659.25254: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10181 1726776659.25287: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10181 1726776659.25316: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10181 1726776659.25342: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10181 1726776659.25387: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10181 1726776659.25413: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10181 1726776659.25439: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10181 1726776659.25477: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10181 1726776659.25490: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10181 1726776659.25585: Evaluated conditional (__kernel_settings_profile_file is defined): True 10181 1726776659.25593: _execute() done 10181 1726776659.25597: dumping result to json 10181 1726776659.25600: done dumping result, returning 10181 1726776659.25605: done running TaskExecutor() for managed_node3/TASK: Check config files [120fa90a-8a95-fee7-0cea-00000000018d] 10181 1726776659.25610: sending task result for task 120fa90a-8a95-fee7-0cea-00000000018d 10181 1726776659.25641: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000018d 10181 1726776659.25645: WORKER PROCESS EXITING 9704 1726776659.25995: no more pending results, returning what we have 9704 1726776659.25999: in VariableManager get_vars() 9704 1726776659.26035: Calling all_inventory to load vars for managed_node3 9704 1726776659.26038: Calling groups_inventory to load vars for managed_node3 9704 1726776659.26040: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.26049: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.26052: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.26055: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.26225: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.26431: done with get_vars() 9704 1726776659.26440: variable 'ansible_search_path' from source: unknown 9704 1726776659.26441: variable 'ansible_search_path' from source: unknown 9704 1726776659.26480: we have included files to process 9704 1726776659.26481: generating all_blocks data 9704 1726776659.26483: done generating all_blocks data 9704 1726776659.26488: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml 9704 1726776659.26489: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml 9704 1726776659.26491: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml for managed_node3 9704 1726776659.27778: done processing included file 9704 1726776659.27781: iterating over new_blocks loaded from include file 9704 1726776659.27782: in VariableManager get_vars() 9704 1726776659.27798: done with get_vars() 9704 1726776659.27799: filtering new block on tags 9704 1726776659.27919: done filtering new block on tags 9704 1726776659.27921: done iterating over new_blocks loaded from include file 9704 1726776659.27922: extending task lists for all hosts with included blocks 9704 1726776659.28259: done extending task lists 9704 1726776659.28260: done processing included files 9704 1726776659.28260: results queue empty 9704 1726776659.28261: checking for any_errors_fatal 9704 1726776659.28264: done checking for any_errors_fatal 9704 1726776659.28264: checking for max_fail_percentage 9704 1726776659.28265: done checking for max_fail_percentage 9704 1726776659.28266: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.28267: done checking to see if all hosts have failed 9704 1726776659.28267: getting the remaining hosts for this loop 9704 1726776659.28269: done getting the remaining hosts for this loop 9704 1726776659.28271: getting the next task for host managed_node3 9704 1726776659.28275: done getting next task for host managed_node3 9704 1726776659.28281: ^ task is: TASK: Create temporary file for expected config 9704 1726776659.28284: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.28286: getting variables 9704 1726776659.28287: in VariableManager get_vars() 9704 1726776659.28297: Calling all_inventory to load vars for managed_node3 9704 1726776659.28299: Calling groups_inventory to load vars for managed_node3 9704 1726776659.28301: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.28306: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.28308: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.28310: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.28458: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.28643: done with get_vars() 9704 1726776659.28651: done getting variables TASK [Create temporary file for expected config] ******************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:2 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.073) 0:00:12.215 **** 9704 1726776659.28711: entering _queue_task() for managed_node3/tempfile 9704 1726776659.28712: Creating lock for tempfile 9704 1726776659.28945: worker is 1 (out of 1 available) 9704 1726776659.28961: exiting _queue_task() for managed_node3/tempfile 9704 1726776659.28972: done queuing things up, now waiting for results queue to drain 9704 1726776659.28974: waiting for pending results... 10184 1726776659.29178: running TaskExecutor() for managed_node3/TASK: Create temporary file for expected config 10184 1726776659.29304: in run() - task 120fa90a-8a95-fee7-0cea-000000000227 10184 1726776659.29321: variable 'ansible_search_path' from source: unknown 10184 1726776659.29326: variable 'ansible_search_path' from source: unknown 10184 1726776659.29359: calling self._execute() 10184 1726776659.29439: variable 'ansible_host' from source: host vars for 'managed_node3' 10184 1726776659.29449: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10184 1726776659.29457: variable 'omit' from source: magic vars 10184 1726776659.29593: variable 'omit' from source: magic vars 10184 1726776659.29639: variable 'omit' from source: magic vars 10184 1726776659.29670: variable 'omit' from source: magic vars 10184 1726776659.29709: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10184 1726776659.29743: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10184 1726776659.29765: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10184 1726776659.29785: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10184 1726776659.29798: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10184 1726776659.29826: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10184 1726776659.29834: variable 'ansible_host' from source: host vars for 'managed_node3' 10184 1726776659.29839: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10184 1726776659.29932: Set connection var ansible_module_compression to ZIP_DEFLATED 10184 1726776659.29942: Set connection var ansible_shell_type to sh 10184 1726776659.29951: Set connection var ansible_timeout to 10 10184 1726776659.29958: Set connection var ansible_pipelining to False 10184 1726776659.29966: Set connection var ansible_shell_executable to /bin/sh 10184 1726776659.29969: Set connection var ansible_connection to ssh 10184 1726776659.29988: variable 'ansible_shell_executable' from source: unknown 10184 1726776659.29994: variable 'ansible_connection' from source: unknown 10184 1726776659.29997: variable 'ansible_module_compression' from source: unknown 10184 1726776659.29999: variable 'ansible_shell_type' from source: unknown 10184 1726776659.30002: variable 'ansible_shell_executable' from source: unknown 10184 1726776659.30004: variable 'ansible_host' from source: host vars for 'managed_node3' 10184 1726776659.30008: variable 'ansible_pipelining' from source: unknown 10184 1726776659.30010: variable 'ansible_timeout' from source: unknown 10184 1726776659.30014: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10184 1726776659.30283: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10184 1726776659.30297: variable 'omit' from source: magic vars 10184 1726776659.30303: starting attempt loop 10184 1726776659.30306: running the handler 10184 1726776659.30318: _low_level_execute_command(): starting 10184 1726776659.30327: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10184 1726776659.33139: stdout chunk (state=2): >>>/root <<< 10184 1726776659.33422: stderr chunk (state=3): >>><<< 10184 1726776659.33436: stdout chunk (state=3): >>><<< 10184 1726776659.33460: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10184 1726776659.33476: _low_level_execute_command(): starting 10184 1726776659.33483: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036 `" && echo ansible-tmp-1726776659.3346972-10184-192176767546036="` echo /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036 `" ) && sleep 0' 10184 1726776659.36232: stdout chunk (state=2): >>>ansible-tmp-1726776659.3346972-10184-192176767546036=/root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036 <<< 10184 1726776659.36357: stderr chunk (state=3): >>><<< 10184 1726776659.36366: stdout chunk (state=3): >>><<< 10184 1726776659.36379: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776659.3346972-10184-192176767546036=/root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036 , stderr= 10184 1726776659.36415: variable 'ansible_module_compression' from source: unknown 10184 1726776659.36455: ANSIBALLZ: Using lock for tempfile 10184 1726776659.36460: ANSIBALLZ: Acquiring lock 10184 1726776659.36462: ANSIBALLZ: Lock acquired: 140007689586304 10184 1726776659.36464: ANSIBALLZ: Creating module 10184 1726776659.45122: ANSIBALLZ: Writing module into payload 10184 1726776659.45176: ANSIBALLZ: Writing module 10184 1726776659.45195: ANSIBALLZ: Renaming module 10184 1726776659.45202: ANSIBALLZ: Done creating module 10184 1726776659.45216: variable 'ansible_facts' from source: unknown 10184 1726776659.45273: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/AnsiballZ_tempfile.py 10184 1726776659.45381: Sending initial data 10184 1726776659.45388: Sent initial data (156 bytes) 10184 1726776659.47973: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp_1_vnh7b /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/AnsiballZ_tempfile.py <<< 10184 1726776659.48919: stderr chunk (state=3): >>><<< 10184 1726776659.48926: stdout chunk (state=3): >>><<< 10184 1726776659.48946: done transferring module to remote 10184 1726776659.48957: _low_level_execute_command(): starting 10184 1726776659.48963: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/ /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/AnsiballZ_tempfile.py && sleep 0' 10184 1726776659.51314: stderr chunk (state=2): >>><<< 10184 1726776659.51321: stdout chunk (state=2): >>><<< 10184 1726776659.51337: _low_level_execute_command() done: rc=0, stdout=, stderr= 10184 1726776659.51342: _low_level_execute_command(): starting 10184 1726776659.51347: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/AnsiballZ_tempfile.py && sleep 0' 10184 1726776659.66380: stdout chunk (state=2): >>> {"changed": true, "path": "/tmp/ansible.a3osdmty.kernel_settings", "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "invocation": {"module_args": {"state": "file", "suffix": ".kernel_settings", "prefix": "ansible.", "path": null}}} <<< 10184 1726776659.67493: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10184 1726776659.67550: stderr chunk (state=3): >>><<< 10184 1726776659.67559: stdout chunk (state=3): >>><<< 10184 1726776659.67575: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "path": "/tmp/ansible.a3osdmty.kernel_settings", "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "invocation": {"module_args": {"state": "file", "suffix": ".kernel_settings", "prefix": "ansible.", "path": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10184 1726776659.67611: done with _execute_module (tempfile, {'state': 'file', 'suffix': '.kernel_settings', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'tempfile', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10184 1726776659.67624: _low_level_execute_command(): starting 10184 1726776659.67633: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776659.3346972-10184-192176767546036/ > /dev/null 2>&1 && sleep 0' 10184 1726776659.70095: stderr chunk (state=2): >>><<< 10184 1726776659.70106: stdout chunk (state=2): >>><<< 10184 1726776659.70120: _low_level_execute_command() done: rc=0, stdout=, stderr= 10184 1726776659.70131: handler run complete 10184 1726776659.70146: attempt loop complete, returning result 10184 1726776659.70150: _execute() done 10184 1726776659.70153: dumping result to json 10184 1726776659.70158: done dumping result, returning 10184 1726776659.70165: done running TaskExecutor() for managed_node3/TASK: Create temporary file for expected config [120fa90a-8a95-fee7-0cea-000000000227] 10184 1726776659.70171: sending task result for task 120fa90a-8a95-fee7-0cea-000000000227 10184 1726776659.70201: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000227 10184 1726776659.70205: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/ansible.a3osdmty.kernel_settings", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 } 9704 1726776659.70365: no more pending results, returning what we have 9704 1726776659.70367: results queue empty 9704 1726776659.70368: checking for any_errors_fatal 9704 1726776659.70370: done checking for any_errors_fatal 9704 1726776659.70370: checking for max_fail_percentage 9704 1726776659.70372: done checking for max_fail_percentage 9704 1726776659.70372: checking to see if all hosts have failed and the running result is not ok 9704 1726776659.70373: done checking to see if all hosts have failed 9704 1726776659.70373: getting the remaining hosts for this loop 9704 1726776659.70374: done getting the remaining hosts for this loop 9704 1726776659.70378: getting the next task for host managed_node3 9704 1726776659.70383: done getting next task for host managed_node3 9704 1726776659.70385: ^ task is: TASK: Put expected contents into temporary file 9704 1726776659.70389: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776659.70392: getting variables 9704 1726776659.70394: in VariableManager get_vars() 9704 1726776659.70465: Calling all_inventory to load vars for managed_node3 9704 1726776659.70468: Calling groups_inventory to load vars for managed_node3 9704 1726776659.70470: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776659.70479: Calling all_plugins_play to load vars for managed_node3 9704 1726776659.70481: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776659.70483: Calling groups_plugins_play to load vars for managed_node3 9704 1726776659.70595: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776659.70710: done with get_vars() 9704 1726776659.70718: done getting variables 9704 1726776659.70761: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Put expected contents into temporary file] ******************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:8 Thursday 19 September 2024 16:10:59 -0400 (0:00:00.420) 0:00:12.636 **** 9704 1726776659.70781: entering _queue_task() for managed_node3/copy 9704 1726776659.70942: worker is 1 (out of 1 available) 9704 1726776659.70955: exiting _queue_task() for managed_node3/copy 9704 1726776659.70966: done queuing things up, now waiting for results queue to drain 9704 1726776659.70968: waiting for pending results... 10217 1726776659.71083: running TaskExecutor() for managed_node3/TASK: Put expected contents into temporary file 10217 1726776659.71188: in run() - task 120fa90a-8a95-fee7-0cea-000000000228 10217 1726776659.71204: variable 'ansible_search_path' from source: unknown 10217 1726776659.71208: variable 'ansible_search_path' from source: unknown 10217 1726776659.71238: calling self._execute() 10217 1726776659.71298: variable 'ansible_host' from source: host vars for 'managed_node3' 10217 1726776659.71306: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10217 1726776659.71315: variable 'omit' from source: magic vars 10217 1726776659.71390: variable 'omit' from source: magic vars 10217 1726776659.71423: variable 'omit' from source: magic vars 10217 1726776659.71657: variable '__kernel_settings_register_profile_conf_tempfile' from source: set_fact 10217 1726776659.71679: variable '__kernel_settings_profile_file' from source: include_vars 10217 1726776659.71734: variable '__kernel_settings_profile_file' from source: include_vars 10217 1726776659.71866: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10217 1726776659.73362: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10217 1726776659.73406: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10217 1726776659.73438: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10217 1726776659.73466: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10217 1726776659.73486: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10217 1726776659.73553: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10217 1726776659.73575: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10217 1726776659.73595: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10217 1726776659.73621: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10217 1726776659.73638: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10217 1726776659.73718: variable 'omit' from source: magic vars 10217 1726776659.73743: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10217 1726776659.73763: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10217 1726776659.73779: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10217 1726776659.73794: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10217 1726776659.73803: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10217 1726776659.73830: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10217 1726776659.73835: variable 'ansible_host' from source: host vars for 'managed_node3' 10217 1726776659.73839: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10217 1726776659.73901: Set connection var ansible_module_compression to ZIP_DEFLATED 10217 1726776659.73909: Set connection var ansible_shell_type to sh 10217 1726776659.73917: Set connection var ansible_timeout to 10 10217 1726776659.73922: Set connection var ansible_pipelining to False 10217 1726776659.73933: Set connection var ansible_shell_executable to /bin/sh 10217 1726776659.73937: Set connection var ansible_connection to ssh 10217 1726776659.73954: variable 'ansible_shell_executable' from source: unknown 10217 1726776659.73958: variable 'ansible_connection' from source: unknown 10217 1726776659.73961: variable 'ansible_module_compression' from source: unknown 10217 1726776659.73964: variable 'ansible_shell_type' from source: unknown 10217 1726776659.73968: variable 'ansible_shell_executable' from source: unknown 10217 1726776659.73971: variable 'ansible_host' from source: host vars for 'managed_node3' 10217 1726776659.73975: variable 'ansible_pipelining' from source: unknown 10217 1726776659.73978: variable 'ansible_timeout' from source: unknown 10217 1726776659.73982: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10217 1726776659.74047: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10217 1726776659.74058: variable 'omit' from source: magic vars 10217 1726776659.74065: starting attempt loop 10217 1726776659.74068: running the handler 10217 1726776659.74081: _low_level_execute_command(): starting 10217 1726776659.74090: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10217 1726776659.76386: stdout chunk (state=2): >>>/root <<< 10217 1726776659.76508: stderr chunk (state=3): >>><<< 10217 1726776659.76515: stdout chunk (state=3): >>><<< 10217 1726776659.76539: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10217 1726776659.76551: _low_level_execute_command(): starting 10217 1726776659.76557: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044 `" && echo ansible-tmp-1726776659.7654696-10217-202917341728044="` echo /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044 `" ) && sleep 0' 10217 1726776659.79031: stdout chunk (state=2): >>>ansible-tmp-1726776659.7654696-10217-202917341728044=/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044 <<< 10217 1726776659.79159: stderr chunk (state=3): >>><<< 10217 1726776659.79165: stdout chunk (state=3): >>><<< 10217 1726776659.79181: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776659.7654696-10217-202917341728044=/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044 , stderr= 10217 1726776659.79249: variable 'ansible_module_compression' from source: unknown 10217 1726776659.79289: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10217 1726776659.79316: variable 'ansible_facts' from source: unknown 10217 1726776659.79384: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_stat.py 10217 1726776659.79469: Sending initial data 10217 1726776659.79476: Sent initial data (152 bytes) 10217 1726776659.81993: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpgcb4iw23 /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_stat.py <<< 10217 1726776659.82967: stderr chunk (state=3): >>><<< 10217 1726776659.82975: stdout chunk (state=3): >>><<< 10217 1726776659.82995: done transferring module to remote 10217 1726776659.83006: _low_level_execute_command(): starting 10217 1726776659.83011: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/ /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_stat.py && sleep 0' 10217 1726776659.85385: stderr chunk (state=2): >>><<< 10217 1726776659.85397: stdout chunk (state=2): >>><<< 10217 1726776659.85413: _low_level_execute_command() done: rc=0, stdout=, stderr= 10217 1726776659.85417: _low_level_execute_command(): starting 10217 1726776659.85423: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_stat.py && sleep 0' 10217 1726776660.01216: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/tmp/ansible.a3osdmty.kernel_settings", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 736921, "dev": 51713, "nlink": 1, "atime": 1726776659.6612563, "mtime": 1726776659.6612563, "ctime": 1726776659.6612563, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "mimetype": "inode/x-empty", "charset": "binary", "version": "2699399081", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/tmp/ansible.a3osdmty.kernel_settings", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10217 1726776660.02387: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10217 1726776660.02443: stderr chunk (state=3): >>><<< 10217 1726776660.02451: stdout chunk (state=3): >>><<< 10217 1726776660.02468: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/tmp/ansible.a3osdmty.kernel_settings", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 0, "inode": 736921, "dev": 51713, "nlink": 1, "atime": 1726776659.6612563, "mtime": 1726776659.6612563, "ctime": 1726776659.6612563, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "mimetype": "inode/x-empty", "charset": "binary", "version": "2699399081", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/tmp/ansible.a3osdmty.kernel_settings", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10217 1726776660.02508: done with _execute_module (ansible.legacy.stat, {'path': '/tmp/ansible.a3osdmty.kernel_settings', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10217 1726776660.02599: Sending initial data 10217 1726776660.02607: Sent initial data (141 bytes) 10217 1726776660.05143: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpe6e6bq3a /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source <<< 10217 1726776660.05419: stderr chunk (state=3): >>><<< 10217 1726776660.05434: stdout chunk (state=3): >>><<< 10217 1726776660.05457: _low_level_execute_command(): starting 10217 1726776660.05464: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/ /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source && sleep 0' 10217 1726776660.07804: stderr chunk (state=2): >>><<< 10217 1726776660.07816: stdout chunk (state=2): >>><<< 10217 1726776660.07836: _low_level_execute_command() done: rc=0, stdout=, stderr= 10217 1726776660.07857: variable 'ansible_module_compression' from source: unknown 10217 1726776660.07896: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.copy-ZIP_DEFLATED 10217 1726776660.07914: variable 'ansible_facts' from source: unknown 10217 1726776660.07975: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_copy.py 10217 1726776660.08069: Sending initial data 10217 1726776660.08075: Sent initial data (152 bytes) 10217 1726776660.10621: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpo5yerk4o /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_copy.py <<< 10217 1726776660.11635: stderr chunk (state=3): >>><<< 10217 1726776660.11646: stdout chunk (state=3): >>><<< 10217 1726776660.11667: done transferring module to remote 10217 1726776660.11677: _low_level_execute_command(): starting 10217 1726776660.11682: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/ /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_copy.py && sleep 0' 10217 1726776660.14077: stderr chunk (state=2): >>><<< 10217 1726776660.14089: stdout chunk (state=2): >>><<< 10217 1726776660.14106: _low_level_execute_command() done: rc=0, stdout=, stderr= 10217 1726776660.14110: _low_level_execute_command(): starting 10217 1726776660.14116: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/AnsiballZ_copy.py && sleep 0' 10217 1726776660.30663: stdout chunk (state=2): >>> {"dest": "/tmp/ansible.a3osdmty.kernel_settings", "src": "/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source", "md5sum": "b82bbe3a22d81e7a442595a59b2d0944", "checksum": "169e46a0442e2493edda7ef1d37dfc572d06525c", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 299, "invocation": {"module_args": {"dest": "/tmp/ansible.a3osdmty.kernel_settings", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source", "_original_basename": "tmpe6e6bq3a", "follow": false, "checksum": "169e46a0442e2493edda7ef1d37dfc572d06525c", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10217 1726776660.31982: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10217 1726776660.31996: stdout chunk (state=3): >>><<< 10217 1726776660.32009: stderr chunk (state=3): >>><<< 10217 1726776660.32024: _low_level_execute_command() done: rc=0, stdout= {"dest": "/tmp/ansible.a3osdmty.kernel_settings", "src": "/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source", "md5sum": "b82bbe3a22d81e7a442595a59b2d0944", "checksum": "169e46a0442e2493edda7ef1d37dfc572d06525c", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 299, "invocation": {"module_args": {"dest": "/tmp/ansible.a3osdmty.kernel_settings", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source", "_original_basename": "tmpe6e6bq3a", "follow": false, "checksum": "169e46a0442e2493edda7ef1d37dfc572d06525c", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10217 1726776660.32067: done with _execute_module (ansible.legacy.copy, {'dest': '/tmp/ansible.a3osdmty.kernel_settings', 'mode': '0600', 'src': '/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source', '_original_basename': 'tmpe6e6bq3a', 'follow': False, 'checksum': '169e46a0442e2493edda7ef1d37dfc572d06525c', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10217 1726776660.32080: _low_level_execute_command(): starting 10217 1726776660.32087: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/ > /dev/null 2>&1 && sleep 0' 10217 1726776660.34762: stderr chunk (state=2): >>><<< 10217 1726776660.34774: stdout chunk (state=2): >>><<< 10217 1726776660.34793: _low_level_execute_command() done: rc=0, stdout=, stderr= 10217 1726776660.34802: handler run complete 10217 1726776660.34832: attempt loop complete, returning result 10217 1726776660.34839: _execute() done 10217 1726776660.34843: dumping result to json 10217 1726776660.34848: done dumping result, returning 10217 1726776660.34855: done running TaskExecutor() for managed_node3/TASK: Put expected contents into temporary file [120fa90a-8a95-fee7-0cea-000000000228] 10217 1726776660.34862: sending task result for task 120fa90a-8a95-fee7-0cea-000000000228 10217 1726776660.34902: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000228 10217 1726776660.34906: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "169e46a0442e2493edda7ef1d37dfc572d06525c", "dest": "/tmp/ansible.a3osdmty.kernel_settings", "gid": 0, "group": "root", "md5sum": "b82bbe3a22d81e7a442595a59b2d0944", "mode": "0600", "owner": "root", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 299, "src": "/root/.ansible/tmp/ansible-tmp-1726776659.7654696-10217-202917341728044/source", "state": "file", "uid": 0 } 9704 1726776660.35374: no more pending results, returning what we have 9704 1726776660.35378: results queue empty 9704 1726776660.35378: checking for any_errors_fatal 9704 1726776660.35386: done checking for any_errors_fatal 9704 1726776660.35387: checking for max_fail_percentage 9704 1726776660.35388: done checking for max_fail_percentage 9704 1726776660.35389: checking to see if all hosts have failed and the running result is not ok 9704 1726776660.35390: done checking to see if all hosts have failed 9704 1726776660.35390: getting the remaining hosts for this loop 9704 1726776660.35391: done getting the remaining hosts for this loop 9704 1726776660.35395: getting the next task for host managed_node3 9704 1726776660.35400: done getting next task for host managed_node3 9704 1726776660.35403: ^ task is: TASK: Ensure python command exists for tests below 9704 1726776660.35406: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776660.35410: getting variables 9704 1726776660.35411: in VariableManager get_vars() 9704 1726776660.35447: Calling all_inventory to load vars for managed_node3 9704 1726776660.35450: Calling groups_inventory to load vars for managed_node3 9704 1726776660.35452: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776660.35461: Calling all_plugins_play to load vars for managed_node3 9704 1726776660.35464: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776660.35467: Calling groups_plugins_play to load vars for managed_node3 9704 1726776660.35638: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776660.35876: done with get_vars() 9704 1726776660.35886: done getting variables 9704 1726776660.35949: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Ensure python command exists for tests below] **************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:14 Thursday 19 September 2024 16:11:00 -0400 (0:00:00.651) 0:00:13.288 **** 9704 1726776660.35979: entering _queue_task() for managed_node3/package 9704 1726776660.36218: worker is 1 (out of 1 available) 9704 1726776660.36235: exiting _queue_task() for managed_node3/package 9704 1726776660.36247: done queuing things up, now waiting for results queue to drain 9704 1726776660.36250: waiting for pending results... 10251 1726776660.36472: running TaskExecutor() for managed_node3/TASK: Ensure python command exists for tests below 10251 1726776660.36611: in run() - task 120fa90a-8a95-fee7-0cea-000000000229 10251 1726776660.36633: variable 'ansible_search_path' from source: unknown 10251 1726776660.36638: variable 'ansible_search_path' from source: unknown 10251 1726776660.36670: calling self._execute() 10251 1726776660.36751: variable 'ansible_host' from source: host vars for 'managed_node3' 10251 1726776660.36761: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10251 1726776660.36771: variable 'omit' from source: magic vars 10251 1726776660.36870: variable 'omit' from source: magic vars 10251 1726776660.36916: variable 'omit' from source: magic vars 10251 1726776660.36948: variable '__kernel_settings_test_python_pkgs' from source: include_vars 10251 1726776660.37232: variable '__kernel_settings_test_python_pkgs' from source: include_vars 10251 1726776660.37507: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10251 1726776660.39835: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10251 1726776660.39912: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10251 1726776660.39953: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10251 1726776660.39987: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10251 1726776660.40013: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10251 1726776660.40110: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10251 1726776660.40142: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10251 1726776660.40167: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10251 1726776660.40206: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10251 1726776660.40220: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10251 1726776660.40332: variable '__kernel_settings_is_ostree' from source: set_fact 10251 1726776660.40340: variable 'omit' from source: magic vars 10251 1726776660.40372: variable 'omit' from source: magic vars 10251 1726776660.40398: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10251 1726776660.40424: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10251 1726776660.40447: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10251 1726776660.40464: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10251 1726776660.40474: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10251 1726776660.40503: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10251 1726776660.40509: variable 'ansible_host' from source: host vars for 'managed_node3' 10251 1726776660.40514: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10251 1726776660.40615: Set connection var ansible_module_compression to ZIP_DEFLATED 10251 1726776660.40623: Set connection var ansible_shell_type to sh 10251 1726776660.40680: Set connection var ansible_timeout to 10 10251 1726776660.40688: Set connection var ansible_pipelining to False 10251 1726776660.40696: Set connection var ansible_shell_executable to /bin/sh 10251 1726776660.40699: Set connection var ansible_connection to ssh 10251 1726776660.40722: variable 'ansible_shell_executable' from source: unknown 10251 1726776660.40731: variable 'ansible_connection' from source: unknown 10251 1726776660.40736: variable 'ansible_module_compression' from source: unknown 10251 1726776660.40739: variable 'ansible_shell_type' from source: unknown 10251 1726776660.40742: variable 'ansible_shell_executable' from source: unknown 10251 1726776660.40744: variable 'ansible_host' from source: host vars for 'managed_node3' 10251 1726776660.40748: variable 'ansible_pipelining' from source: unknown 10251 1726776660.40750: variable 'ansible_timeout' from source: unknown 10251 1726776660.40753: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10251 1726776660.40843: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10251 1726776660.40855: variable 'omit' from source: magic vars 10251 1726776660.40862: starting attempt loop 10251 1726776660.40865: running the handler 10251 1726776660.40954: variable 'ansible_facts' from source: unknown 10251 1726776660.41075: _low_level_execute_command(): starting 10251 1726776660.41085: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10251 1726776660.43614: stdout chunk (state=2): >>>/root <<< 10251 1726776660.43766: stderr chunk (state=3): >>><<< 10251 1726776660.43776: stdout chunk (state=3): >>><<< 10251 1726776660.43799: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10251 1726776660.43814: _low_level_execute_command(): starting 10251 1726776660.43822: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575 `" && echo ansible-tmp-1726776660.438086-10251-19455344672575="` echo /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575 `" ) && sleep 0' 10251 1726776660.46548: stdout chunk (state=2): >>>ansible-tmp-1726776660.438086-10251-19455344672575=/root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575 <<< 10251 1726776660.46701: stderr chunk (state=3): >>><<< 10251 1726776660.46710: stdout chunk (state=3): >>><<< 10251 1726776660.46732: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776660.438086-10251-19455344672575=/root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575 , stderr= 10251 1726776660.46761: variable 'ansible_module_compression' from source: unknown 10251 1726776660.46820: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 10251 1726776660.46866: variable 'ansible_facts' from source: unknown 10251 1726776660.47009: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/AnsiballZ_dnf.py 10251 1726776660.47478: Sending initial data 10251 1726776660.47485: Sent initial data (149 bytes) 10251 1726776660.50295: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpmg_iwcby /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/AnsiballZ_dnf.py <<< 10251 1726776660.52036: stderr chunk (state=3): >>><<< 10251 1726776660.52047: stdout chunk (state=3): >>><<< 10251 1726776660.52070: done transferring module to remote 10251 1726776660.52087: _low_level_execute_command(): starting 10251 1726776660.52094: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/ /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/AnsiballZ_dnf.py && sleep 0' 10251 1726776660.54767: stderr chunk (state=2): >>><<< 10251 1726776660.54778: stdout chunk (state=2): >>><<< 10251 1726776660.54795: _low_level_execute_command() done: rc=0, stdout=, stderr= 10251 1726776660.54801: _low_level_execute_command(): starting 10251 1726776660.54808: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/AnsiballZ_dnf.py && sleep 0' 10251 1726776663.13341: stdout chunk (state=2): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} <<< 10251 1726776663.22535: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10251 1726776663.22585: stderr chunk (state=3): >>><<< 10251 1726776663.22594: stdout chunk (state=3): >>><<< 10251 1726776663.22619: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10251 1726776663.22657: done with _execute_module (ansible.legacy.dnf, {'name': ['python3-configobj'], 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10251 1726776663.22668: _low_level_execute_command(): starting 10251 1726776663.22674: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776660.438086-10251-19455344672575/ > /dev/null 2>&1 && sleep 0' 10251 1726776663.25186: stderr chunk (state=2): >>><<< 10251 1726776663.25196: stdout chunk (state=2): >>><<< 10251 1726776663.25210: _low_level_execute_command() done: rc=0, stdout=, stderr= 10251 1726776663.25217: handler run complete 10251 1726776663.25246: attempt loop complete, returning result 10251 1726776663.25250: _execute() done 10251 1726776663.25255: dumping result to json 10251 1726776663.25262: done dumping result, returning 10251 1726776663.25268: done running TaskExecutor() for managed_node3/TASK: Ensure python command exists for tests below [120fa90a-8a95-fee7-0cea-000000000229] 10251 1726776663.25274: sending task result for task 120fa90a-8a95-fee7-0cea-000000000229 10251 1726776663.25304: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000229 10251 1726776663.25308: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 9704 1726776663.25466: no more pending results, returning what we have 9704 1726776663.25469: results queue empty 9704 1726776663.25469: checking for any_errors_fatal 9704 1726776663.25477: done checking for any_errors_fatal 9704 1726776663.25477: checking for max_fail_percentage 9704 1726776663.25479: done checking for max_fail_percentage 9704 1726776663.25479: checking to see if all hosts have failed and the running result is not ok 9704 1726776663.25480: done checking to see if all hosts have failed 9704 1726776663.25480: getting the remaining hosts for this loop 9704 1726776663.25482: done getting the remaining hosts for this loop 9704 1726776663.25485: getting the next task for host managed_node3 9704 1726776663.25490: done getting next task for host managed_node3 9704 1726776663.25492: ^ task is: TASK: Diff expected vs actual content 9704 1726776663.25495: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776663.25498: getting variables 9704 1726776663.25500: in VariableManager get_vars() 9704 1726776663.25533: Calling all_inventory to load vars for managed_node3 9704 1726776663.25536: Calling groups_inventory to load vars for managed_node3 9704 1726776663.25539: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776663.25547: Calling all_plugins_play to load vars for managed_node3 9704 1726776663.25549: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776663.25552: Calling groups_plugins_play to load vars for managed_node3 9704 1726776663.25667: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776663.25782: done with get_vars() 9704 1726776663.25790: done getting variables 9704 1726776663.25834: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Diff expected vs actual content] ***************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:20 Thursday 19 September 2024 16:11:03 -0400 (0:00:02.898) 0:00:16.187 **** 9704 1726776663.25859: entering _queue_task() for managed_node3/shell 9704 1726776663.26022: worker is 1 (out of 1 available) 9704 1726776663.26042: exiting _queue_task() for managed_node3/shell 9704 1726776663.26053: done queuing things up, now waiting for results queue to drain 9704 1726776663.26056: waiting for pending results... 10305 1726776663.26165: running TaskExecutor() for managed_node3/TASK: Diff expected vs actual content 10305 1726776663.26269: in run() - task 120fa90a-8a95-fee7-0cea-00000000022a 10305 1726776663.26284: variable 'ansible_search_path' from source: unknown 10305 1726776663.26288: variable 'ansible_search_path' from source: unknown 10305 1726776663.26316: calling self._execute() 10305 1726776663.26380: variable 'ansible_host' from source: host vars for 'managed_node3' 10305 1726776663.26388: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10305 1726776663.26398: variable 'omit' from source: magic vars 10305 1726776663.26470: variable 'omit' from source: magic vars 10305 1726776663.26503: variable 'omit' from source: magic vars 10305 1726776663.26764: variable '__kernel_settings_test_python_cmd' from source: include_vars 10305 1726776663.26775: variable '__kernel_settings_register_profile_conf_tempfile' from source: set_fact 10305 1726776663.26785: variable '__kernel_settings_profile_filename' from source: role '' exported vars 10305 1726776663.26841: variable '__kernel_settings_profile_dir' from source: role '' exported vars 10305 1726776663.26899: variable '__kernel_settings_profile_parent' from source: set_fact 10305 1726776663.26906: variable '__kernel_settings_tuned_profile' from source: role '' exported vars 10305 1726776663.26985: variable 'omit' from source: magic vars 10305 1726776663.27019: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10305 1726776663.27047: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10305 1726776663.27066: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10305 1726776663.27079: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10305 1726776663.27089: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10305 1726776663.27113: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10305 1726776663.27118: variable 'ansible_host' from source: host vars for 'managed_node3' 10305 1726776663.27122: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10305 1726776663.27189: Set connection var ansible_module_compression to ZIP_DEFLATED 10305 1726776663.27197: Set connection var ansible_shell_type to sh 10305 1726776663.27205: Set connection var ansible_timeout to 10 10305 1726776663.27210: Set connection var ansible_pipelining to False 10305 1726776663.27217: Set connection var ansible_shell_executable to /bin/sh 10305 1726776663.27220: Set connection var ansible_connection to ssh 10305 1726776663.27238: variable 'ansible_shell_executable' from source: unknown 10305 1726776663.27242: variable 'ansible_connection' from source: unknown 10305 1726776663.27246: variable 'ansible_module_compression' from source: unknown 10305 1726776663.27249: variable 'ansible_shell_type' from source: unknown 10305 1726776663.27253: variable 'ansible_shell_executable' from source: unknown 10305 1726776663.27256: variable 'ansible_host' from source: host vars for 'managed_node3' 10305 1726776663.27260: variable 'ansible_pipelining' from source: unknown 10305 1726776663.27264: variable 'ansible_timeout' from source: unknown 10305 1726776663.27268: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10305 1726776663.27357: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10305 1726776663.27368: variable 'omit' from source: magic vars 10305 1726776663.27374: starting attempt loop 10305 1726776663.27378: running the handler 10305 1726776663.27384: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10305 1726776663.27397: _low_level_execute_command(): starting 10305 1726776663.27403: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10305 1726776663.29750: stdout chunk (state=2): >>>/root <<< 10305 1726776663.29872: stderr chunk (state=3): >>><<< 10305 1726776663.29880: stdout chunk (state=3): >>><<< 10305 1726776663.29900: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10305 1726776663.29914: _low_level_execute_command(): starting 10305 1726776663.29920: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323 `" && echo ansible-tmp-1726776663.2990816-10305-142154923108323="` echo /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323 `" ) && sleep 0' 10305 1726776663.32442: stdout chunk (state=2): >>>ansible-tmp-1726776663.2990816-10305-142154923108323=/root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323 <<< 10305 1726776663.32573: stderr chunk (state=3): >>><<< 10305 1726776663.32580: stdout chunk (state=3): >>><<< 10305 1726776663.32595: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776663.2990816-10305-142154923108323=/root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323 , stderr= 10305 1726776663.32619: variable 'ansible_module_compression' from source: unknown 10305 1726776663.32659: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 10305 1726776663.32687: variable 'ansible_facts' from source: unknown 10305 1726776663.32760: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/AnsiballZ_command.py 10305 1726776663.32859: Sending initial data 10305 1726776663.32866: Sent initial data (155 bytes) 10305 1726776663.35411: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpblo5_on8 /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/AnsiballZ_command.py <<< 10305 1726776663.36376: stderr chunk (state=3): >>><<< 10305 1726776663.36385: stdout chunk (state=3): >>><<< 10305 1726776663.36405: done transferring module to remote 10305 1726776663.36415: _low_level_execute_command(): starting 10305 1726776663.36420: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/ /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/AnsiballZ_command.py && sleep 0' 10305 1726776663.38866: stderr chunk (state=2): >>><<< 10305 1726776663.38875: stdout chunk (state=2): >>><<< 10305 1726776663.38891: _low_level_execute_command() done: rc=0, stdout=, stderr= 10305 1726776663.38897: _low_level_execute_command(): starting 10305 1726776663.38905: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/AnsiballZ_command.py && sleep 0' 10305 1726776663.57680: stdout chunk (state=2): >>> {"changed": true, "stdout": "", "stderr": "", "rc": 0, "cmd": "/usr/libexec/platform-python -c 'import sys\nfrom configobj import ConfigObj\nc1 = ConfigObj(sys.argv[1])\nc2 = ConfigObj(sys.argv[2])\nsys.exit(not c1 == c2)' /tmp/ansible.a3osdmty.kernel_settings /etc/tuned/kernel_settings/tuned.conf\n", "start": "2024-09-19 16:11:03.544180", "end": "2024-09-19 16:11:03.575408", "delta": "0:00:00.031228", "msg": "", "invocation": {"module_args": {"_raw_params": "/usr/libexec/platform-python -c 'import sys\nfrom configobj import ConfigObj\nc1 = ConfigObj(sys.argv[1])\nc2 = ConfigObj(sys.argv[2])\nsys.exit(not c1 == c2)' /tmp/ansible.a3osdmty.kernel_settings /etc/tuned/kernel_settings/tuned.conf\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 10305 1726776663.58912: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10305 1726776663.58961: stderr chunk (state=3): >>><<< 10305 1726776663.58970: stdout chunk (state=3): >>><<< 10305 1726776663.58988: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "", "rc": 0, "cmd": "/usr/libexec/platform-python -c 'import sys\nfrom configobj import ConfigObj\nc1 = ConfigObj(sys.argv[1])\nc2 = ConfigObj(sys.argv[2])\nsys.exit(not c1 == c2)' /tmp/ansible.a3osdmty.kernel_settings /etc/tuned/kernel_settings/tuned.conf\n", "start": "2024-09-19 16:11:03.544180", "end": "2024-09-19 16:11:03.575408", "delta": "0:00:00.031228", "msg": "", "invocation": {"module_args": {"_raw_params": "/usr/libexec/platform-python -c 'import sys\nfrom configobj import ConfigObj\nc1 = ConfigObj(sys.argv[1])\nc2 = ConfigObj(sys.argv[2])\nsys.exit(not c1 == c2)' /tmp/ansible.a3osdmty.kernel_settings /etc/tuned/kernel_settings/tuned.conf\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10305 1726776663.59016: done with _execute_module (ansible.legacy.command, {'_raw_params': "/usr/libexec/platform-python -c 'import sys\nfrom configobj import ConfigObj\nc1 = ConfigObj(sys.argv[1])\nc2 = ConfigObj(sys.argv[2])\nsys.exit(not c1 == c2)' /tmp/ansible.a3osdmty.kernel_settings /etc/tuned/kernel_settings/tuned.conf\n", '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10305 1726776663.59025: _low_level_execute_command(): starting 10305 1726776663.59032: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776663.2990816-10305-142154923108323/ > /dev/null 2>&1 && sleep 0' 10305 1726776663.61457: stderr chunk (state=2): >>><<< 10305 1726776663.61464: stdout chunk (state=2): >>><<< 10305 1726776663.61477: _low_level_execute_command() done: rc=0, stdout=, stderr= 10305 1726776663.61484: handler run complete 10305 1726776663.61500: Evaluated conditional (False): False 10305 1726776663.61509: attempt loop complete, returning result 10305 1726776663.61513: _execute() done 10305 1726776663.61517: dumping result to json 10305 1726776663.61522: done dumping result, returning 10305 1726776663.61530: done running TaskExecutor() for managed_node3/TASK: Diff expected vs actual content [120fa90a-8a95-fee7-0cea-00000000022a] 10305 1726776663.61540: sending task result for task 120fa90a-8a95-fee7-0cea-00000000022a 10305 1726776663.61572: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000022a 10305 1726776663.61576: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "cmd": "/usr/libexec/platform-python -c 'import sys\nfrom configobj import ConfigObj\nc1 = ConfigObj(sys.argv[1])\nc2 = ConfigObj(sys.argv[2])\nsys.exit(not c1 == c2)' /tmp/ansible.a3osdmty.kernel_settings /etc/tuned/kernel_settings/tuned.conf\n", "delta": "0:00:00.031228", "end": "2024-09-19 16:11:03.575408", "rc": 0, "start": "2024-09-19 16:11:03.544180" } 9704 1726776663.61706: no more pending results, returning what we have 9704 1726776663.61709: results queue empty 9704 1726776663.61710: checking for any_errors_fatal 9704 1726776663.61715: done checking for any_errors_fatal 9704 1726776663.61716: checking for max_fail_percentage 9704 1726776663.61717: done checking for max_fail_percentage 9704 1726776663.61718: checking to see if all hosts have failed and the running result is not ok 9704 1726776663.61719: done checking to see if all hosts have failed 9704 1726776663.61719: getting the remaining hosts for this loop 9704 1726776663.61720: done getting the remaining hosts for this loop 9704 1726776663.61723: getting the next task for host managed_node3 9704 1726776663.61731: done getting next task for host managed_node3 9704 1726776663.61732: ^ task is: TASK: Verify expected content 9704 1726776663.61735: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776663.61739: getting variables 9704 1726776663.61740: in VariableManager get_vars() 9704 1726776663.61770: Calling all_inventory to load vars for managed_node3 9704 1726776663.61773: Calling groups_inventory to load vars for managed_node3 9704 1726776663.61775: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776663.61783: Calling all_plugins_play to load vars for managed_node3 9704 1726776663.61786: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776663.61789: Calling groups_plugins_play to load vars for managed_node3 9704 1726776663.61907: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776663.62059: done with get_vars() 9704 1726776663.62066: done getting variables 9704 1726776663.62106: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Verify expected content] ************************************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:33 Thursday 19 September 2024 16:11:03 -0400 (0:00:00.362) 0:00:16.550 **** 9704 1726776663.62125: entering _queue_task() for managed_node3/set_fact 9704 1726776663.62286: worker is 1 (out of 1 available) 9704 1726776663.62300: exiting _queue_task() for managed_node3/set_fact 9704 1726776663.62311: done queuing things up, now waiting for results queue to drain 9704 1726776663.62313: waiting for pending results... 10321 1726776663.62424: running TaskExecutor() for managed_node3/TASK: Verify expected content 10321 1726776663.62527: in run() - task 120fa90a-8a95-fee7-0cea-00000000022b 10321 1726776663.62547: variable 'ansible_search_path' from source: unknown 10321 1726776663.62551: variable 'ansible_search_path' from source: unknown 10321 1726776663.62576: calling self._execute() 10321 1726776663.62639: variable 'ansible_host' from source: host vars for 'managed_node3' 10321 1726776663.62645: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10321 1726776663.62650: variable 'omit' from source: magic vars 10321 1726776663.62969: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10321 1726776663.63146: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10321 1726776663.63179: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10321 1726776663.63206: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10321 1726776663.63235: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10321 1726776663.63308: variable '__kernel_settings_register_profile_conf_result' from source: set_fact 10321 1726776663.63334: Evaluated conditional (__kernel_settings_register_profile_conf_result is failed): False 10321 1726776663.63339: when evaluation is False, skipping this task 10321 1726776663.63342: _execute() done 10321 1726776663.63345: dumping result to json 10321 1726776663.63349: done dumping result, returning 10321 1726776663.63354: done running TaskExecutor() for managed_node3/TASK: Verify expected content [120fa90a-8a95-fee7-0cea-00000000022b] 10321 1726776663.63360: sending task result for task 120fa90a-8a95-fee7-0cea-00000000022b 10321 1726776663.63383: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000022b 10321 1726776663.63387: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_profile_conf_result is failed", "skip_reason": "Conditional result was False" } 9704 1726776663.63485: no more pending results, returning what we have 9704 1726776663.63488: results queue empty 9704 1726776663.63488: checking for any_errors_fatal 9704 1726776663.63493: done checking for any_errors_fatal 9704 1726776663.63493: checking for max_fail_percentage 9704 1726776663.63495: done checking for max_fail_percentage 9704 1726776663.63495: checking to see if all hosts have failed and the running result is not ok 9704 1726776663.63496: done checking to see if all hosts have failed 9704 1726776663.63496: getting the remaining hosts for this loop 9704 1726776663.63497: done getting the remaining hosts for this loop 9704 1726776663.63500: getting the next task for host managed_node3 9704 1726776663.63506: done getting next task for host managed_node3 9704 1726776663.63508: ^ task is: TASK: Show diff - may not reflect actual ConfigObj differences 9704 1726776663.63511: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776663.63514: getting variables 9704 1726776663.63515: in VariableManager get_vars() 9704 1726776663.63543: Calling all_inventory to load vars for managed_node3 9704 1726776663.63546: Calling groups_inventory to load vars for managed_node3 9704 1726776663.63548: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776663.63555: Calling all_plugins_play to load vars for managed_node3 9704 1726776663.63559: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776663.63561: Calling groups_plugins_play to load vars for managed_node3 9704 1726776663.63662: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776663.63774: done with get_vars() 9704 1726776663.63781: done getting variables 9704 1726776663.63819: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show diff - may not reflect actual ConfigObj differences] **************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:38 Thursday 19 September 2024 16:11:03 -0400 (0:00:00.017) 0:00:16.567 **** 9704 1726776663.63842: entering _queue_task() for managed_node3/command 9704 1726776663.63985: worker is 1 (out of 1 available) 9704 1726776663.63999: exiting _queue_task() for managed_node3/command 9704 1726776663.64010: done queuing things up, now waiting for results queue to drain 9704 1726776663.64012: waiting for pending results... 10322 1726776663.64116: running TaskExecutor() for managed_node3/TASK: Show diff - may not reflect actual ConfigObj differences 10322 1726776663.64215: in run() - task 120fa90a-8a95-fee7-0cea-00000000022c 10322 1726776663.64233: variable 'ansible_search_path' from source: unknown 10322 1726776663.64237: variable 'ansible_search_path' from source: unknown 10322 1726776663.64264: calling self._execute() 10322 1726776663.64320: variable 'ansible_host' from source: host vars for 'managed_node3' 10322 1726776663.64329: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10322 1726776663.64340: variable 'omit' from source: magic vars 10322 1726776663.64647: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10322 1726776663.64865: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10322 1726776663.64898: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10322 1726776663.64922: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10322 1726776663.64951: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10322 1726776663.65024: variable '__kernel_settings_register_profile_conf_result' from source: set_fact 10322 1726776663.65047: Evaluated conditional (__kernel_settings_register_profile_conf_result is failed): False 10322 1726776663.65052: when evaluation is False, skipping this task 10322 1726776663.65056: _execute() done 10322 1726776663.65059: dumping result to json 10322 1726776663.65063: done dumping result, returning 10322 1726776663.65068: done running TaskExecutor() for managed_node3/TASK: Show diff - may not reflect actual ConfigObj differences [120fa90a-8a95-fee7-0cea-00000000022c] 10322 1726776663.65074: sending task result for task 120fa90a-8a95-fee7-0cea-00000000022c 10322 1726776663.65093: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000022c 10322 1726776663.65095: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_profile_conf_result is failed", "skip_reason": "Conditional result was False" } 9704 1726776663.65242: no more pending results, returning what we have 9704 1726776663.65245: results queue empty 9704 1726776663.65246: checking for any_errors_fatal 9704 1726776663.65250: done checking for any_errors_fatal 9704 1726776663.65250: checking for max_fail_percentage 9704 1726776663.65252: done checking for max_fail_percentage 9704 1726776663.65252: checking to see if all hosts have failed and the running result is not ok 9704 1726776663.65253: done checking to see if all hosts have failed 9704 1726776663.65253: getting the remaining hosts for this loop 9704 1726776663.65255: done getting the remaining hosts for this loop 9704 1726776663.65257: getting the next task for host managed_node3 9704 1726776663.65262: done getting next task for host managed_node3 9704 1726776663.65264: ^ task is: TASK: Get active_profile file 9704 1726776663.65267: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776663.65270: getting variables 9704 1726776663.65271: in VariableManager get_vars() 9704 1726776663.65295: Calling all_inventory to load vars for managed_node3 9704 1726776663.65296: Calling groups_inventory to load vars for managed_node3 9704 1726776663.65297: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776663.65303: Calling all_plugins_play to load vars for managed_node3 9704 1726776663.65305: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776663.65306: Calling groups_plugins_play to load vars for managed_node3 9704 1726776663.65406: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776663.65559: done with get_vars() 9704 1726776663.65565: done getting variables TASK [Get active_profile file] ************************************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:46 Thursday 19 September 2024 16:11:03 -0400 (0:00:00.017) 0:00:16.585 **** 9704 1726776663.65623: entering _queue_task() for managed_node3/slurp 9704 1726776663.65765: worker is 1 (out of 1 available) 9704 1726776663.65779: exiting _queue_task() for managed_node3/slurp 9704 1726776663.65791: done queuing things up, now waiting for results queue to drain 9704 1726776663.65792: waiting for pending results... 10323 1726776663.65898: running TaskExecutor() for managed_node3/TASK: Get active_profile file 10323 1726776663.65992: in run() - task 120fa90a-8a95-fee7-0cea-00000000022d 10323 1726776663.66007: variable 'ansible_search_path' from source: unknown 10323 1726776663.66011: variable 'ansible_search_path' from source: unknown 10323 1726776663.66041: calling self._execute() 10323 1726776663.66102: variable 'ansible_host' from source: host vars for 'managed_node3' 10323 1726776663.66110: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10323 1726776663.66119: variable 'omit' from source: magic vars 10323 1726776663.66192: variable 'omit' from source: magic vars 10323 1726776663.66224: variable 'omit' from source: magic vars 10323 1726776663.66250: variable 'omit' from source: magic vars 10323 1726776663.66281: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10323 1726776663.66307: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10323 1726776663.66326: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10323 1726776663.66344: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10323 1726776663.66353: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10323 1726776663.66373: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10323 1726776663.66376: variable 'ansible_host' from source: host vars for 'managed_node3' 10323 1726776663.66379: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10323 1726776663.66444: Set connection var ansible_module_compression to ZIP_DEFLATED 10323 1726776663.66450: Set connection var ansible_shell_type to sh 10323 1726776663.66456: Set connection var ansible_timeout to 10 10323 1726776663.66460: Set connection var ansible_pipelining to False 10323 1726776663.66464: Set connection var ansible_shell_executable to /bin/sh 10323 1726776663.66466: Set connection var ansible_connection to ssh 10323 1726776663.66478: variable 'ansible_shell_executable' from source: unknown 10323 1726776663.66481: variable 'ansible_connection' from source: unknown 10323 1726776663.66483: variable 'ansible_module_compression' from source: unknown 10323 1726776663.66485: variable 'ansible_shell_type' from source: unknown 10323 1726776663.66487: variable 'ansible_shell_executable' from source: unknown 10323 1726776663.66489: variable 'ansible_host' from source: host vars for 'managed_node3' 10323 1726776663.66492: variable 'ansible_pipelining' from source: unknown 10323 1726776663.66495: variable 'ansible_timeout' from source: unknown 10323 1726776663.66497: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10323 1726776663.66640: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10323 1726776663.66652: variable 'omit' from source: magic vars 10323 1726776663.66657: starting attempt loop 10323 1726776663.66661: running the handler 10323 1726776663.66671: _low_level_execute_command(): starting 10323 1726776663.66679: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10323 1726776663.69045: stdout chunk (state=2): >>>/root <<< 10323 1726776663.69165: stderr chunk (state=3): >>><<< 10323 1726776663.69171: stdout chunk (state=3): >>><<< 10323 1726776663.69188: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10323 1726776663.69200: _low_level_execute_command(): starting 10323 1726776663.69206: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272 `" && echo ansible-tmp-1726776663.691952-10323-67618131274272="` echo /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272 `" ) && sleep 0' 10323 1726776663.71695: stdout chunk (state=2): >>>ansible-tmp-1726776663.691952-10323-67618131274272=/root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272 <<< 10323 1726776663.71825: stderr chunk (state=3): >>><<< 10323 1726776663.71833: stdout chunk (state=3): >>><<< 10323 1726776663.71845: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776663.691952-10323-67618131274272=/root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272 , stderr= 10323 1726776663.71875: variable 'ansible_module_compression' from source: unknown 10323 1726776663.71912: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.slurp-ZIP_DEFLATED 10323 1726776663.71937: variable 'ansible_facts' from source: unknown 10323 1726776663.72005: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/AnsiballZ_slurp.py 10323 1726776663.72101: Sending initial data 10323 1726776663.72108: Sent initial data (151 bytes) 10323 1726776663.74645: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp_36krrna /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/AnsiballZ_slurp.py <<< 10323 1726776663.75585: stderr chunk (state=3): >>><<< 10323 1726776663.75591: stdout chunk (state=3): >>><<< 10323 1726776663.75607: done transferring module to remote 10323 1726776663.75617: _low_level_execute_command(): starting 10323 1726776663.75622: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/ /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/AnsiballZ_slurp.py && sleep 0' 10323 1726776663.78169: stderr chunk (state=2): >>><<< 10323 1726776663.78175: stdout chunk (state=2): >>><<< 10323 1726776663.78188: _low_level_execute_command() done: rc=0, stdout=, stderr= 10323 1726776663.78192: _low_level_execute_command(): starting 10323 1726776663.78197: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/AnsiballZ_slurp.py && sleep 0' 10323 1726776663.94125: stdout chunk (state=2): >>> {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"src": "/etc/tuned/active_profile"}}} <<< 10323 1726776663.95227: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10323 1726776663.95277: stderr chunk (state=3): >>><<< 10323 1726776663.95284: stdout chunk (state=3): >>><<< 10323 1726776663.95299: _low_level_execute_command() done: rc=0, stdout= {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"src": "/etc/tuned/active_profile"}}} , stderr=Shared connection to 10.31.8.186 closed. 10323 1726776663.95330: done with _execute_module (slurp, {'src': '/etc/tuned/active_profile', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10323 1726776663.95343: _low_level_execute_command(): starting 10323 1726776663.95349: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776663.691952-10323-67618131274272/ > /dev/null 2>&1 && sleep 0' 10323 1726776663.97789: stderr chunk (state=2): >>><<< 10323 1726776663.97797: stdout chunk (state=2): >>><<< 10323 1726776663.97810: _low_level_execute_command() done: rc=0, stdout=, stderr= 10323 1726776663.97817: handler run complete 10323 1726776663.97832: attempt loop complete, returning result 10323 1726776663.97837: _execute() done 10323 1726776663.97841: dumping result to json 10323 1726776663.97845: done dumping result, returning 10323 1726776663.97851: done running TaskExecutor() for managed_node3/TASK: Get active_profile file [120fa90a-8a95-fee7-0cea-00000000022d] 10323 1726776663.97858: sending task result for task 120fa90a-8a95-fee7-0cea-00000000022d 10323 1726776663.97888: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000022d 10323 1726776663.97892: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "encoding": "base64", "source": "/etc/tuned/active_profile" } 9704 1726776663.98079: no more pending results, returning what we have 9704 1726776663.98083: results queue empty 9704 1726776663.98084: checking for any_errors_fatal 9704 1726776663.98089: done checking for any_errors_fatal 9704 1726776663.98089: checking for max_fail_percentage 9704 1726776663.98091: done checking for max_fail_percentage 9704 1726776663.98092: checking to see if all hosts have failed and the running result is not ok 9704 1726776663.98092: done checking to see if all hosts have failed 9704 1726776663.98093: getting the remaining hosts for this loop 9704 1726776663.98094: done getting the remaining hosts for this loop 9704 1726776663.98097: getting the next task for host managed_node3 9704 1726776663.98102: done getting next task for host managed_node3 9704 1726776663.98104: ^ task is: TASK: Check that active_profile ends with kernel_settings 9704 1726776663.98107: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776663.98110: getting variables 9704 1726776663.98111: in VariableManager get_vars() 9704 1726776663.98139: Calling all_inventory to load vars for managed_node3 9704 1726776663.98142: Calling groups_inventory to load vars for managed_node3 9704 1726776663.98143: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776663.98149: Calling all_plugins_play to load vars for managed_node3 9704 1726776663.98151: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776663.98153: Calling groups_plugins_play to load vars for managed_node3 9704 1726776663.98260: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776663.98377: done with get_vars() 9704 1726776663.98385: done getting variables 9704 1726776663.98426: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check that active_profile ends with kernel_settings] ********************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:53 Thursday 19 September 2024 16:11:03 -0400 (0:00:00.328) 0:00:16.913 **** 9704 1726776663.98451: entering _queue_task() for managed_node3/set_fact 9704 1726776663.98612: worker is 1 (out of 1 available) 9704 1726776663.98626: exiting _queue_task() for managed_node3/set_fact 9704 1726776663.98642: done queuing things up, now waiting for results queue to drain 9704 1726776663.98644: waiting for pending results... 10342 1726776663.98755: running TaskExecutor() for managed_node3/TASK: Check that active_profile ends with kernel_settings 10342 1726776663.98858: in run() - task 120fa90a-8a95-fee7-0cea-00000000022e 10342 1726776663.98872: variable 'ansible_search_path' from source: unknown 10342 1726776663.98875: variable 'ansible_search_path' from source: unknown 10342 1726776663.98899: calling self._execute() 10342 1726776663.98963: variable 'ansible_host' from source: host vars for 'managed_node3' 10342 1726776663.98971: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10342 1726776663.98979: variable 'omit' from source: magic vars 10342 1726776663.99324: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10342 1726776664.00887: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10342 1726776664.00936: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10342 1726776664.00964: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10342 1726776664.00991: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10342 1726776664.01011: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10342 1726776664.01066: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10342 1726776664.01363: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10342 1726776664.01391: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10342 1726776664.01414: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10342 1726776664.01439: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10342 1726776664.01526: variable '__kernel_settings_register_profile_conf_active_profile' from source: set_fact 10342 1726776664.01568: Evaluated conditional (__kernel_settings_register_profile_conf_active_profile.content | b64decode is not search('(^| )kernel_settings$')): False 10342 1726776664.01573: when evaluation is False, skipping this task 10342 1726776664.01577: _execute() done 10342 1726776664.01580: dumping result to json 10342 1726776664.01584: done dumping result, returning 10342 1726776664.01590: done running TaskExecutor() for managed_node3/TASK: Check that active_profile ends with kernel_settings [120fa90a-8a95-fee7-0cea-00000000022e] 10342 1726776664.01595: sending task result for task 120fa90a-8a95-fee7-0cea-00000000022e 10342 1726776664.01617: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000022e 10342 1726776664.01620: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_profile_conf_active_profile.content | b64decode is not search('(^| )kernel_settings$')", "skip_reason": "Conditional result was False" } 9704 1726776664.01718: no more pending results, returning what we have 9704 1726776664.01722: results queue empty 9704 1726776664.01723: checking for any_errors_fatal 9704 1726776664.01730: done checking for any_errors_fatal 9704 1726776664.01730: checking for max_fail_percentage 9704 1726776664.01732: done checking for max_fail_percentage 9704 1726776664.01732: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.01733: done checking to see if all hosts have failed 9704 1726776664.01733: getting the remaining hosts for this loop 9704 1726776664.01735: done getting the remaining hosts for this loop 9704 1726776664.01738: getting the next task for host managed_node3 9704 1726776664.01743: done getting next task for host managed_node3 9704 1726776664.01745: ^ task is: TASK: Get profile_mode file 9704 1726776664.01748: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.01751: getting variables 9704 1726776664.01752: in VariableManager get_vars() 9704 1726776664.01780: Calling all_inventory to load vars for managed_node3 9704 1726776664.01782: Calling groups_inventory to load vars for managed_node3 9704 1726776664.01784: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.01798: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.01801: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.01804: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.01959: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.02071: done with get_vars() 9704 1726776664.02077: done getting variables TASK [Get profile_mode file] *************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:60 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.036) 0:00:16.950 **** 9704 1726776664.02137: entering _queue_task() for managed_node3/slurp 9704 1726776664.02289: worker is 1 (out of 1 available) 9704 1726776664.02303: exiting _queue_task() for managed_node3/slurp 9704 1726776664.02313: done queuing things up, now waiting for results queue to drain 9704 1726776664.02315: waiting for pending results... 10343 1726776664.02424: running TaskExecutor() for managed_node3/TASK: Get profile_mode file 10343 1726776664.02524: in run() - task 120fa90a-8a95-fee7-0cea-00000000022f 10343 1726776664.02542: variable 'ansible_search_path' from source: unknown 10343 1726776664.02547: variable 'ansible_search_path' from source: unknown 10343 1726776664.02571: calling self._execute() 10343 1726776664.02633: variable 'ansible_host' from source: host vars for 'managed_node3' 10343 1726776664.02639: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10343 1726776664.02646: variable 'omit' from source: magic vars 10343 1726776664.02712: variable 'omit' from source: magic vars 10343 1726776664.02745: variable 'omit' from source: magic vars 10343 1726776664.02766: variable 'omit' from source: magic vars 10343 1726776664.02795: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10343 1726776664.02818: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10343 1726776664.02837: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10343 1726776664.02849: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10343 1726776664.02857: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10343 1726776664.02879: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10343 1726776664.02883: variable 'ansible_host' from source: host vars for 'managed_node3' 10343 1726776664.02885: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10343 1726776664.02950: Set connection var ansible_module_compression to ZIP_DEFLATED 10343 1726776664.02956: Set connection var ansible_shell_type to sh 10343 1726776664.02961: Set connection var ansible_timeout to 10 10343 1726776664.02965: Set connection var ansible_pipelining to False 10343 1726776664.02970: Set connection var ansible_shell_executable to /bin/sh 10343 1726776664.02972: Set connection var ansible_connection to ssh 10343 1726776664.02986: variable 'ansible_shell_executable' from source: unknown 10343 1726776664.02990: variable 'ansible_connection' from source: unknown 10343 1726776664.02992: variable 'ansible_module_compression' from source: unknown 10343 1726776664.02994: variable 'ansible_shell_type' from source: unknown 10343 1726776664.02996: variable 'ansible_shell_executable' from source: unknown 10343 1726776664.02997: variable 'ansible_host' from source: host vars for 'managed_node3' 10343 1726776664.03000: variable 'ansible_pipelining' from source: unknown 10343 1726776664.03001: variable 'ansible_timeout' from source: unknown 10343 1726776664.03003: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10343 1726776664.03138: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10343 1726776664.03151: variable 'omit' from source: magic vars 10343 1726776664.03157: starting attempt loop 10343 1726776664.03161: running the handler 10343 1726776664.03173: _low_level_execute_command(): starting 10343 1726776664.03181: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10343 1726776664.05530: stdout chunk (state=2): >>>/root <<< 10343 1726776664.05651: stderr chunk (state=3): >>><<< 10343 1726776664.05658: stdout chunk (state=3): >>><<< 10343 1726776664.05677: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10343 1726776664.05692: _low_level_execute_command(): starting 10343 1726776664.05698: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408 `" && echo ansible-tmp-1726776664.0568597-10343-217572397742408="` echo /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408 `" ) && sleep 0' 10343 1726776664.08203: stdout chunk (state=2): >>>ansible-tmp-1726776664.0568597-10343-217572397742408=/root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408 <<< 10343 1726776664.08335: stderr chunk (state=3): >>><<< 10343 1726776664.08341: stdout chunk (state=3): >>><<< 10343 1726776664.08356: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776664.0568597-10343-217572397742408=/root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408 , stderr= 10343 1726776664.08390: variable 'ansible_module_compression' from source: unknown 10343 1726776664.08433: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.slurp-ZIP_DEFLATED 10343 1726776664.08468: variable 'ansible_facts' from source: unknown 10343 1726776664.08560: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/AnsiballZ_slurp.py 10343 1726776664.09025: Sending initial data 10343 1726776664.09036: Sent initial data (153 bytes) 10343 1726776664.11689: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpk1s_0lqo /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/AnsiballZ_slurp.py <<< 10343 1726776664.13536: stderr chunk (state=3): >>><<< 10343 1726776664.13545: stdout chunk (state=3): >>><<< 10343 1726776664.13570: done transferring module to remote 10343 1726776664.13584: _low_level_execute_command(): starting 10343 1726776664.13590: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/ /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/AnsiballZ_slurp.py && sleep 0' 10343 1726776664.16150: stderr chunk (state=2): >>><<< 10343 1726776664.16158: stdout chunk (state=2): >>><<< 10343 1726776664.16173: _low_level_execute_command() done: rc=0, stdout=, stderr= 10343 1726776664.16177: _low_level_execute_command(): starting 10343 1726776664.16182: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/AnsiballZ_slurp.py && sleep 0' 10343 1726776664.31500: stdout chunk (state=2): >>> {"content": "bWFudWFsCg==", "source": "/etc/tuned/profile_mode", "encoding": "base64", "invocation": {"module_args": {"src": "/etc/tuned/profile_mode"}}} <<< 10343 1726776664.32578: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10343 1726776664.32627: stderr chunk (state=3): >>><<< 10343 1726776664.32636: stdout chunk (state=3): >>><<< 10343 1726776664.32651: _low_level_execute_command() done: rc=0, stdout= {"content": "bWFudWFsCg==", "source": "/etc/tuned/profile_mode", "encoding": "base64", "invocation": {"module_args": {"src": "/etc/tuned/profile_mode"}}} , stderr=Shared connection to 10.31.8.186 closed. 10343 1726776664.32682: done with _execute_module (slurp, {'src': '/etc/tuned/profile_mode', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10343 1726776664.32694: _low_level_execute_command(): starting 10343 1726776664.32700: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776664.0568597-10343-217572397742408/ > /dev/null 2>&1 && sleep 0' 10343 1726776664.35124: stderr chunk (state=2): >>><<< 10343 1726776664.35136: stdout chunk (state=2): >>><<< 10343 1726776664.35151: _low_level_execute_command() done: rc=0, stdout=, stderr= 10343 1726776664.35158: handler run complete 10343 1726776664.35170: attempt loop complete, returning result 10343 1726776664.35174: _execute() done 10343 1726776664.35177: dumping result to json 10343 1726776664.35181: done dumping result, returning 10343 1726776664.35188: done running TaskExecutor() for managed_node3/TASK: Get profile_mode file [120fa90a-8a95-fee7-0cea-00000000022f] 10343 1726776664.35195: sending task result for task 120fa90a-8a95-fee7-0cea-00000000022f 10343 1726776664.35223: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000022f 10343 1726776664.35227: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "bWFudWFsCg==", "encoding": "base64", "source": "/etc/tuned/profile_mode" } 9704 1726776664.35344: no more pending results, returning what we have 9704 1726776664.35347: results queue empty 9704 1726776664.35348: checking for any_errors_fatal 9704 1726776664.35353: done checking for any_errors_fatal 9704 1726776664.35354: checking for max_fail_percentage 9704 1726776664.35355: done checking for max_fail_percentage 9704 1726776664.35356: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.35356: done checking to see if all hosts have failed 9704 1726776664.35357: getting the remaining hosts for this loop 9704 1726776664.35358: done getting the remaining hosts for this loop 9704 1726776664.35361: getting the next task for host managed_node3 9704 1726776664.35366: done getting next task for host managed_node3 9704 1726776664.35368: ^ task is: TASK: Check that profile_mode is manual 9704 1726776664.35371: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.35374: getting variables 9704 1726776664.35375: in VariableManager get_vars() 9704 1726776664.35405: Calling all_inventory to load vars for managed_node3 9704 1726776664.35408: Calling groups_inventory to load vars for managed_node3 9704 1726776664.35410: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.35418: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.35421: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.35423: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.35552: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.35668: done with get_vars() 9704 1726776664.35676: done getting variables 9704 1726776664.35717: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check that profile_mode is manual] *************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:67 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.336) 0:00:17.286 **** 9704 1726776664.35741: entering _queue_task() for managed_node3/set_fact 9704 1726776664.35906: worker is 1 (out of 1 available) 9704 1726776664.35921: exiting _queue_task() for managed_node3/set_fact 9704 1726776664.35934: done queuing things up, now waiting for results queue to drain 9704 1726776664.35936: waiting for pending results... 10359 1726776664.36051: running TaskExecutor() for managed_node3/TASK: Check that profile_mode is manual 10359 1726776664.36156: in run() - task 120fa90a-8a95-fee7-0cea-000000000230 10359 1726776664.36172: variable 'ansible_search_path' from source: unknown 10359 1726776664.36176: variable 'ansible_search_path' from source: unknown 10359 1726776664.36204: calling self._execute() 10359 1726776664.36271: variable 'ansible_host' from source: host vars for 'managed_node3' 10359 1726776664.36279: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10359 1726776664.36288: variable 'omit' from source: magic vars 10359 1726776664.36643: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10359 1726776664.38418: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10359 1726776664.38467: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10359 1726776664.38495: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10359 1726776664.38521: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10359 1726776664.38547: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10359 1726776664.38600: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10359 1726776664.38719: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10359 1726776664.38751: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10359 1726776664.38775: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10359 1726776664.38796: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10359 1726776664.38889: variable '__kernel_settings_register_profile_conf_profile_mode' from source: set_fact 10359 1726776664.38923: Evaluated conditional (__kernel_settings_register_profile_conf_profile_mode.content | b64decode is not match('manual')): False 10359 1726776664.38928: when evaluation is False, skipping this task 10359 1726776664.38936: _execute() done 10359 1726776664.38941: dumping result to json 10359 1726776664.38944: done dumping result, returning 10359 1726776664.38950: done running TaskExecutor() for managed_node3/TASK: Check that profile_mode is manual [120fa90a-8a95-fee7-0cea-000000000230] 10359 1726776664.38955: sending task result for task 120fa90a-8a95-fee7-0cea-000000000230 10359 1726776664.38978: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000230 10359 1726776664.38981: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_profile_conf_profile_mode.content | b64decode is not match('manual')", "skip_reason": "Conditional result was False" } 9704 1726776664.39139: no more pending results, returning what we have 9704 1726776664.39142: results queue empty 9704 1726776664.39143: checking for any_errors_fatal 9704 1726776664.39148: done checking for any_errors_fatal 9704 1726776664.39149: checking for max_fail_percentage 9704 1726776664.39150: done checking for max_fail_percentage 9704 1726776664.39151: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.39151: done checking to see if all hosts have failed 9704 1726776664.39152: getting the remaining hosts for this loop 9704 1726776664.39153: done getting the remaining hosts for this loop 9704 1726776664.39157: getting the next task for host managed_node3 9704 1726776664.39162: done getting next task for host managed_node3 9704 1726776664.39164: ^ task is: TASK: Get the bootloader specific config file 9704 1726776664.39167: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.39170: getting variables 9704 1726776664.39172: in VariableManager get_vars() 9704 1726776664.39206: Calling all_inventory to load vars for managed_node3 9704 1726776664.39210: Calling groups_inventory to load vars for managed_node3 9704 1726776664.39212: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.39220: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.39222: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.39224: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.39368: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.39482: done with get_vars() 9704 1726776664.39488: done getting variables 9704 1726776664.39527: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Get the bootloader specific config file] ********************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:74 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.038) 0:00:17.324 **** 9704 1726776664.39551: entering _queue_task() for managed_node3/command 9704 1726776664.39711: worker is 1 (out of 1 available) 9704 1726776664.39726: exiting _queue_task() for managed_node3/command 9704 1726776664.39740: done queuing things up, now waiting for results queue to drain 9704 1726776664.39742: waiting for pending results... 10360 1726776664.39859: running TaskExecutor() for managed_node3/TASK: Get the bootloader specific config file 10360 1726776664.39962: in run() - task 120fa90a-8a95-fee7-0cea-000000000231 10360 1726776664.39978: variable 'ansible_search_path' from source: unknown 10360 1726776664.39982: variable 'ansible_search_path' from source: unknown 10360 1726776664.40008: calling self._execute() 10360 1726776664.40071: variable 'ansible_host' from source: host vars for 'managed_node3' 10360 1726776664.40078: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10360 1726776664.40083: variable 'omit' from source: magic vars 10360 1726776664.40400: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10360 1726776664.42101: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10360 1726776664.42148: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10360 1726776664.42176: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10360 1726776664.42212: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10360 1726776664.42237: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10360 1726776664.42290: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10360 1726776664.42311: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10360 1726776664.42334: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10360 1726776664.42364: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10360 1726776664.42376: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10360 1726776664.42457: Evaluated conditional (__kernel_settings_blcmdline_value | d()): False 10360 1726776664.42462: when evaluation is False, skipping this task 10360 1726776664.42466: _execute() done 10360 1726776664.42470: dumping result to json 10360 1726776664.42473: done dumping result, returning 10360 1726776664.42478: done running TaskExecutor() for managed_node3/TASK: Get the bootloader specific config file [120fa90a-8a95-fee7-0cea-000000000231] 10360 1726776664.42484: sending task result for task 120fa90a-8a95-fee7-0cea-000000000231 10360 1726776664.42506: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000231 10360 1726776664.42509: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_blcmdline_value | d()", "skip_reason": "Conditional result was False" } 9704 1726776664.42608: no more pending results, returning what we have 9704 1726776664.42611: results queue empty 9704 1726776664.42611: checking for any_errors_fatal 9704 1726776664.42616: done checking for any_errors_fatal 9704 1726776664.42616: checking for max_fail_percentage 9704 1726776664.42618: done checking for max_fail_percentage 9704 1726776664.42618: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.42619: done checking to see if all hosts have failed 9704 1726776664.42619: getting the remaining hosts for this loop 9704 1726776664.42620: done getting the remaining hosts for this loop 9704 1726776664.42623: getting the next task for host managed_node3 9704 1726776664.42630: done getting next task for host managed_node3 9704 1726776664.42632: ^ task is: TASK: Verify bootloader settings value 9704 1726776664.42635: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.42638: getting variables 9704 1726776664.42639: in VariableManager get_vars() 9704 1726776664.42666: Calling all_inventory to load vars for managed_node3 9704 1726776664.42669: Calling groups_inventory to load vars for managed_node3 9704 1726776664.42670: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.42678: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.42680: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.42682: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.42842: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.43025: done with get_vars() 9704 1726776664.43036: done getting variables 9704 1726776664.43092: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Verify bootloader settings value] **************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:85 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.035) 0:00:17.359 **** 9704 1726776664.43122: entering _queue_task() for managed_node3/set_fact 9704 1726776664.43310: worker is 1 (out of 1 available) 9704 1726776664.43325: exiting _queue_task() for managed_node3/set_fact 9704 1726776664.43337: done queuing things up, now waiting for results queue to drain 9704 1726776664.43338: waiting for pending results... 10362 1726776664.43538: running TaskExecutor() for managed_node3/TASK: Verify bootloader settings value 10362 1726776664.43660: in run() - task 120fa90a-8a95-fee7-0cea-000000000232 10362 1726776664.43676: variable 'ansible_search_path' from source: unknown 10362 1726776664.43680: variable 'ansible_search_path' from source: unknown 10362 1726776664.43705: calling self._execute() 10362 1726776664.43770: variable 'ansible_host' from source: host vars for 'managed_node3' 10362 1726776664.43779: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10362 1726776664.43789: variable 'omit' from source: magic vars 10362 1726776664.44094: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10362 1726776664.45825: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10362 1726776664.45879: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10362 1726776664.45911: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10362 1726776664.45944: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10362 1726776664.45967: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10362 1726776664.46037: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10362 1726776664.46064: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10362 1726776664.46088: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10362 1726776664.46127: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10362 1726776664.46144: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10362 1726776664.46253: Evaluated conditional (__kernel_settings_blcmdline_value | d()): False 10362 1726776664.46259: when evaluation is False, skipping this task 10362 1726776664.46262: _execute() done 10362 1726776664.46265: dumping result to json 10362 1726776664.46268: done dumping result, returning 10362 1726776664.46273: done running TaskExecutor() for managed_node3/TASK: Verify bootloader settings value [120fa90a-8a95-fee7-0cea-000000000232] 10362 1726776664.46278: sending task result for task 120fa90a-8a95-fee7-0cea-000000000232 10362 1726776664.46305: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000232 10362 1726776664.46309: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_blcmdline_value | d()", "skip_reason": "Conditional result was False" } 9704 1726776664.46766: no more pending results, returning what we have 9704 1726776664.46769: results queue empty 9704 1726776664.46770: checking for any_errors_fatal 9704 1726776664.46774: done checking for any_errors_fatal 9704 1726776664.46775: checking for max_fail_percentage 9704 1726776664.46776: done checking for max_fail_percentage 9704 1726776664.46777: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.46777: done checking to see if all hosts have failed 9704 1726776664.46778: getting the remaining hosts for this loop 9704 1726776664.46779: done getting the remaining hosts for this loop 9704 1726776664.46782: getting the next task for host managed_node3 9704 1726776664.46789: done getting next task for host managed_node3 9704 1726776664.46791: ^ task is: TASK: Check if kernel_settings_reboot_required is set if needed 9704 1726776664.46793: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.46796: getting variables 9704 1726776664.46797: in VariableManager get_vars() 9704 1726776664.46826: Calling all_inventory to load vars for managed_node3 9704 1726776664.46831: Calling groups_inventory to load vars for managed_node3 9704 1726776664.46833: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.46842: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.46845: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.46848: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.47072: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.47261: done with get_vars() 9704 1726776664.47271: done getting variables 9704 1726776664.47327: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check if kernel_settings_reboot_required is set if needed] *************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:49 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.042) 0:00:17.402 **** 9704 1726776664.47357: entering _queue_task() for managed_node3/set_fact 9704 1726776664.47558: worker is 1 (out of 1 available) 9704 1726776664.47571: exiting _queue_task() for managed_node3/set_fact 9704 1726776664.47583: done queuing things up, now waiting for results queue to drain 9704 1726776664.47585: waiting for pending results... 10364 1726776664.47795: running TaskExecutor() for managed_node3/TASK: Check if kernel_settings_reboot_required is set if needed 10364 1726776664.47915: in run() - task 120fa90a-8a95-fee7-0cea-00000000018e 10364 1726776664.47933: variable 'ansible_search_path' from source: unknown 10364 1726776664.47939: variable 'ansible_search_path' from source: unknown 10364 1726776664.47973: calling self._execute() 10364 1726776664.48056: variable 'ansible_host' from source: host vars for 'managed_node3' 10364 1726776664.48065: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10364 1726776664.48074: variable 'omit' from source: magic vars 10364 1726776664.48487: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10364 1726776664.50334: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10364 1726776664.50380: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10364 1726776664.50417: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10364 1726776664.50448: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10364 1726776664.50470: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10364 1726776664.50524: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10364 1726776664.50548: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10364 1726776664.50569: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10364 1726776664.50597: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10364 1726776664.50608: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10364 1726776664.50694: Evaluated conditional (__kernel_settings_check_reboot | d(false)): False 10364 1726776664.50699: when evaluation is False, skipping this task 10364 1726776664.50703: _execute() done 10364 1726776664.50706: dumping result to json 10364 1726776664.50710: done dumping result, returning 10364 1726776664.50716: done running TaskExecutor() for managed_node3/TASK: Check if kernel_settings_reboot_required is set if needed [120fa90a-8a95-fee7-0cea-00000000018e] 10364 1726776664.50721: sending task result for task 120fa90a-8a95-fee7-0cea-00000000018e 10364 1726776664.50747: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000018e 10364 1726776664.50751: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_check_reboot | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776664.50857: no more pending results, returning what we have 9704 1726776664.50860: results queue empty 9704 1726776664.50861: checking for any_errors_fatal 9704 1726776664.50865: done checking for any_errors_fatal 9704 1726776664.50866: checking for max_fail_percentage 9704 1726776664.50867: done checking for max_fail_percentage 9704 1726776664.50868: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.50869: done checking to see if all hosts have failed 9704 1726776664.50869: getting the remaining hosts for this loop 9704 1726776664.50870: done getting the remaining hosts for this loop 9704 1726776664.50873: getting the next task for host managed_node3 9704 1726776664.50878: done getting next task for host managed_node3 9704 1726776664.50880: ^ task is: TASK: Assert success 9704 1726776664.50882: ^ state is: HOST STATE: block=2, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.50885: getting variables 9704 1726776664.50886: in VariableManager get_vars() 9704 1726776664.50914: Calling all_inventory to load vars for managed_node3 9704 1726776664.50917: Calling groups_inventory to load vars for managed_node3 9704 1726776664.50918: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.50928: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.50931: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.50937: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.51070: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.51186: done with get_vars() 9704 1726776664.51193: done getting variables 9704 1726776664.51265: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=False, class_only=True) TASK [Assert success] ********************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings.yml:56 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.039) 0:00:17.441 **** 9704 1726776664.51286: entering _queue_task() for managed_node3/assert 9704 1726776664.51287: Creating lock for assert 9704 1726776664.51463: worker is 1 (out of 1 available) 9704 1726776664.51479: exiting _queue_task() for managed_node3/assert 9704 1726776664.51489: done queuing things up, now waiting for results queue to drain 9704 1726776664.51492: waiting for pending results... 10366 1726776664.51603: running TaskExecutor() for managed_node3/TASK: Assert success 10366 1726776664.51713: in run() - task 120fa90a-8a95-fee7-0cea-00000000018f 10366 1726776664.51731: variable 'ansible_search_path' from source: unknown 10366 1726776664.51736: variable 'ansible_search_path' from source: unknown 10366 1726776664.51766: calling self._execute() 10366 1726776664.51831: variable 'ansible_host' from source: host vars for 'managed_node3' 10366 1726776664.51839: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10366 1726776664.51848: variable 'omit' from source: magic vars 10366 1726776664.51920: variable 'omit' from source: magic vars 10366 1726776664.51951: variable 'omit' from source: magic vars 10366 1726776664.51973: variable 'omit' from source: magic vars 10366 1726776664.52003: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10366 1726776664.52027: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10366 1726776664.52053: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10366 1726776664.52070: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10366 1726776664.52081: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10366 1726776664.52107: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10366 1726776664.52112: variable 'ansible_host' from source: host vars for 'managed_node3' 10366 1726776664.52116: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10366 1726776664.52210: Set connection var ansible_module_compression to ZIP_DEFLATED 10366 1726776664.52219: Set connection var ansible_shell_type to sh 10366 1726776664.52227: Set connection var ansible_timeout to 10 10366 1726776664.52235: Set connection var ansible_pipelining to False 10366 1726776664.52243: Set connection var ansible_shell_executable to /bin/sh 10366 1726776664.52247: Set connection var ansible_connection to ssh 10366 1726776664.52264: variable 'ansible_shell_executable' from source: unknown 10366 1726776664.52269: variable 'ansible_connection' from source: unknown 10366 1726776664.52274: variable 'ansible_module_compression' from source: unknown 10366 1726776664.52277: variable 'ansible_shell_type' from source: unknown 10366 1726776664.52280: variable 'ansible_shell_executable' from source: unknown 10366 1726776664.52283: variable 'ansible_host' from source: host vars for 'managed_node3' 10366 1726776664.52286: variable 'ansible_pipelining' from source: unknown 10366 1726776664.52288: variable 'ansible_timeout' from source: unknown 10366 1726776664.52290: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10366 1726776664.52390: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10366 1726776664.52402: variable 'omit' from source: magic vars 10366 1726776664.52408: starting attempt loop 10366 1726776664.52412: running the handler 10366 1726776664.53031: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10366 1726776664.54543: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10366 1726776664.54590: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10366 1726776664.54617: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10366 1726776664.54646: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10366 1726776664.54677: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10366 1726776664.54726: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10366 1726776664.54750: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10366 1726776664.54768: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10366 1726776664.54796: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10366 1726776664.54808: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10366 1726776664.54884: variable '__kernel_settings_success' from source: set_fact 10366 1726776664.54901: Evaluated conditional (__kernel_settings_success | d(false)): True 10366 1726776664.54908: handler run complete 10366 1726776664.54925: attempt loop complete, returning result 10366 1726776664.54930: _execute() done 10366 1726776664.54934: dumping result to json 10366 1726776664.54937: done dumping result, returning 10366 1726776664.54944: done running TaskExecutor() for managed_node3/TASK: Assert success [120fa90a-8a95-fee7-0cea-00000000018f] 10366 1726776664.54949: sending task result for task 120fa90a-8a95-fee7-0cea-00000000018f 10366 1726776664.54972: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000018f 10366 1726776664.54976: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false } MSG: All assertions passed 9704 1726776664.55083: no more pending results, returning what we have 9704 1726776664.55085: results queue empty 9704 1726776664.55086: checking for any_errors_fatal 9704 1726776664.55090: done checking for any_errors_fatal 9704 1726776664.55090: checking for max_fail_percentage 9704 1726776664.55092: done checking for max_fail_percentage 9704 1726776664.55092: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.55093: done checking to see if all hosts have failed 9704 1726776664.55093: getting the remaining hosts for this loop 9704 1726776664.55094: done getting the remaining hosts for this loop 9704 1726776664.55097: getting the next task for host managed_node3 9704 1726776664.55103: done getting next task for host managed_node3 9704 1726776664.55107: ^ task is: TASK: Check ansible_managed, fingerprint in generated files 9704 1726776664.55108: ^ state is: HOST STATE: block=2, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.55111: getting variables 9704 1726776664.55113: in VariableManager get_vars() 9704 1726776664.55146: Calling all_inventory to load vars for managed_node3 9704 1726776664.55148: Calling groups_inventory to load vars for managed_node3 9704 1726776664.55150: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.55158: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.55165: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.55168: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.55555: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.55735: done with get_vars() 9704 1726776664.55743: done getting variables TASK [Check ansible_managed, fingerprint in generated files] ******************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:35 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.045) 0:00:17.487 **** 9704 1726776664.55834: entering _queue_task() for managed_node3/include_tasks 9704 1726776664.56050: worker is 1 (out of 1 available) 9704 1726776664.56064: exiting _queue_task() for managed_node3/include_tasks 9704 1726776664.56075: done queuing things up, now waiting for results queue to drain 9704 1726776664.56077: waiting for pending results... 10369 1726776664.56287: running TaskExecutor() for managed_node3/TASK: Check ansible_managed, fingerprint in generated files 10369 1726776664.56414: in run() - task 120fa90a-8a95-fee7-0cea-00000000000a 10369 1726776664.56437: variable 'ansible_search_path' from source: unknown 10369 1726776664.56471: calling self._execute() 10369 1726776664.56551: variable 'ansible_host' from source: host vars for 'managed_node3' 10369 1726776664.56561: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10369 1726776664.56570: variable 'omit' from source: magic vars 10369 1726776664.56664: _execute() done 10369 1726776664.56670: dumping result to json 10369 1726776664.56674: done dumping result, returning 10369 1726776664.56679: done running TaskExecutor() for managed_node3/TASK: Check ansible_managed, fingerprint in generated files [120fa90a-8a95-fee7-0cea-00000000000a] 10369 1726776664.56686: sending task result for task 120fa90a-8a95-fee7-0cea-00000000000a 10369 1726776664.56717: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000000a 10369 1726776664.56722: WORKER PROCESS EXITING 9704 1726776664.56942: no more pending results, returning what we have 9704 1726776664.56948: in VariableManager get_vars() 9704 1726776664.56973: Calling all_inventory to load vars for managed_node3 9704 1726776664.56975: Calling groups_inventory to load vars for managed_node3 9704 1726776664.56976: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.56982: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.56984: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.56986: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.57096: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.57207: done with get_vars() 9704 1726776664.57212: variable 'ansible_search_path' from source: unknown 9704 1726776664.57221: we have included files to process 9704 1726776664.57222: generating all_blocks data 9704 1726776664.57223: done generating all_blocks data 9704 1726776664.57226: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/check_header.yml 9704 1726776664.57227: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/check_header.yml 9704 1726776664.57232: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/check_header.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/check_header.yml for managed_node3 9704 1726776664.57457: done processing included file 9704 1726776664.57459: iterating over new_blocks loaded from include file 9704 1726776664.57460: in VariableManager get_vars() 9704 1726776664.57470: done with get_vars() 9704 1726776664.57472: filtering new block on tags 9704 1726776664.57490: done filtering new block on tags 9704 1726776664.57492: done iterating over new_blocks loaded from include file 9704 1726776664.57492: extending task lists for all hosts with included blocks 9704 1726776664.58408: done extending task lists 9704 1726776664.58409: done processing included files 9704 1726776664.58409: results queue empty 9704 1726776664.58410: checking for any_errors_fatal 9704 1726776664.58412: done checking for any_errors_fatal 9704 1726776664.58413: checking for max_fail_percentage 9704 1726776664.58413: done checking for max_fail_percentage 9704 1726776664.58414: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.58414: done checking to see if all hosts have failed 9704 1726776664.58415: getting the remaining hosts for this loop 9704 1726776664.58415: done getting the remaining hosts for this loop 9704 1726776664.58417: getting the next task for host managed_node3 9704 1726776664.58419: done getting next task for host managed_node3 9704 1726776664.58421: ^ task is: TASK: Get file 9704 1726776664.58422: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.58424: getting variables 9704 1726776664.58424: in VariableManager get_vars() 9704 1726776664.58435: Calling all_inventory to load vars for managed_node3 9704 1726776664.58436: Calling groups_inventory to load vars for managed_node3 9704 1726776664.58438: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.58442: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.58444: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.58446: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.58543: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.58648: done with get_vars() 9704 1726776664.58656: done getting variables TASK [Get file] **************************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/check_header.yml:3 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.028) 0:00:17.515 **** 9704 1726776664.58702: entering _queue_task() for managed_node3/slurp 9704 1726776664.58906: worker is 1 (out of 1 available) 9704 1726776664.58923: exiting _queue_task() for managed_node3/slurp 9704 1726776664.58939: done queuing things up, now waiting for results queue to drain 9704 1726776664.58942: waiting for pending results... 10371 1726776664.59109: running TaskExecutor() for managed_node3/TASK: Get file 10371 1726776664.59236: in run() - task 120fa90a-8a95-fee7-0cea-00000000028a 10371 1726776664.59253: variable 'ansible_search_path' from source: unknown 10371 1726776664.59260: variable 'ansible_search_path' from source: unknown 10371 1726776664.59293: calling self._execute() 10371 1726776664.59376: variable 'ansible_host' from source: host vars for 'managed_node3' 10371 1726776664.59387: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10371 1726776664.59395: variable 'omit' from source: magic vars 10371 1726776664.59848: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10371 1726776664.60144: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10371 1726776664.60183: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10371 1726776664.60211: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10371 1726776664.60242: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10371 1726776664.60312: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10371 1726776664.60335: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10371 1726776664.60353: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10371 1726776664.60384: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10371 1726776664.60506: Evaluated conditional (not __file_content is defined): True 10371 1726776664.60514: variable 'omit' from source: magic vars 10371 1726776664.60546: variable 'omit' from source: magic vars 10371 1726776664.60568: variable '__file' from source: include params 10371 1726776664.60630: variable '__file' from source: include params 10371 1726776664.60639: variable '__kernel_settings_profile_filename' from source: role '' exported vars 10371 1726776664.60686: variable '__kernel_settings_profile_filename' from source: role '' exported vars 10371 1726776664.60739: variable '__kernel_settings_profile_dir' from source: role '' exported vars 10371 1726776664.60813: variable '__kernel_settings_profile_parent' from source: set_fact 10371 1726776664.60821: variable '__kernel_settings_tuned_profile' from source: role '' exported vars 10371 1726776664.60899: variable 'omit' from source: magic vars 10371 1726776664.60919: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10371 1726776664.60942: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10371 1726776664.60960: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10371 1726776664.60974: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10371 1726776664.60984: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10371 1726776664.61009: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10371 1726776664.61015: variable 'ansible_host' from source: host vars for 'managed_node3' 10371 1726776664.61019: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10371 1726776664.61085: Set connection var ansible_module_compression to ZIP_DEFLATED 10371 1726776664.61093: Set connection var ansible_shell_type to sh 10371 1726776664.61101: Set connection var ansible_timeout to 10 10371 1726776664.61106: Set connection var ansible_pipelining to False 10371 1726776664.61114: Set connection var ansible_shell_executable to /bin/sh 10371 1726776664.61117: Set connection var ansible_connection to ssh 10371 1726776664.61144: variable 'ansible_shell_executable' from source: unknown 10371 1726776664.61149: variable 'ansible_connection' from source: unknown 10371 1726776664.61152: variable 'ansible_module_compression' from source: unknown 10371 1726776664.61155: variable 'ansible_shell_type' from source: unknown 10371 1726776664.61157: variable 'ansible_shell_executable' from source: unknown 10371 1726776664.61160: variable 'ansible_host' from source: host vars for 'managed_node3' 10371 1726776664.61164: variable 'ansible_pipelining' from source: unknown 10371 1726776664.61167: variable 'ansible_timeout' from source: unknown 10371 1726776664.61170: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10371 1726776664.61306: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10371 1726776664.61318: variable 'omit' from source: magic vars 10371 1726776664.61323: starting attempt loop 10371 1726776664.61326: running the handler 10371 1726776664.61340: _low_level_execute_command(): starting 10371 1726776664.61348: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10371 1726776664.63758: stdout chunk (state=2): >>>/root <<< 10371 1726776664.63904: stderr chunk (state=3): >>><<< 10371 1726776664.63911: stdout chunk (state=3): >>><<< 10371 1726776664.63936: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10371 1726776664.63953: _low_level_execute_command(): starting 10371 1726776664.63959: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967 `" && echo ansible-tmp-1726776664.6394653-10371-32665271035967="` echo /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967 `" ) && sleep 0' 10371 1726776664.66631: stdout chunk (state=2): >>>ansible-tmp-1726776664.6394653-10371-32665271035967=/root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967 <<< 10371 1726776664.66762: stderr chunk (state=3): >>><<< 10371 1726776664.66770: stdout chunk (state=3): >>><<< 10371 1726776664.66785: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776664.6394653-10371-32665271035967=/root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967 , stderr= 10371 1726776664.66823: variable 'ansible_module_compression' from source: unknown 10371 1726776664.66860: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.slurp-ZIP_DEFLATED 10371 1726776664.66887: variable 'ansible_facts' from source: unknown 10371 1726776664.66960: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/AnsiballZ_slurp.py 10371 1726776664.67062: Sending initial data 10371 1726776664.67069: Sent initial data (152 bytes) 10371 1726776664.69612: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpjs1zk0ev /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/AnsiballZ_slurp.py <<< 10371 1726776664.70560: stderr chunk (state=3): >>><<< 10371 1726776664.70568: stdout chunk (state=3): >>><<< 10371 1726776664.70586: done transferring module to remote 10371 1726776664.70598: _low_level_execute_command(): starting 10371 1726776664.70604: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/ /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/AnsiballZ_slurp.py && sleep 0' 10371 1726776664.72975: stderr chunk (state=2): >>><<< 10371 1726776664.72983: stdout chunk (state=2): >>><<< 10371 1726776664.72997: _low_level_execute_command() done: rc=0, stdout=, stderr= 10371 1726776664.73001: _low_level_execute_command(): starting 10371 1726776664.73007: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/AnsiballZ_slurp.py && sleep 0' 10371 1726776664.88292: stdout chunk (state=2): >>> {"content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6a2VybmVsX3NldHRpbmdzCgpbbWFpbl0Kc3VtbWFyeSA9IGtlcm5lbCBzZXR0aW5ncwpbc3lzY3RsXQpmcy5lcG9sbC5tYXhfdXNlcl93YXRjaGVzID0gNzg1NTkyCmZzLmZpbGUtbWF4ID0gMzc5NzI0CltzeXNmc10KL3N5cy9rZXJuZWwvZGVidWcveDg2L2licnNfZW5hYmxlZCA9IDAKL3N5cy9rZXJuZWwvZGVidWcveDg2L3B0aV9lbmFibGVkID0gMAovc3lzL2tlcm5lbC9kZWJ1Zy94ODYvcmV0cF9lbmFibGVkID0gMApbdm1dCnRyYW5zcGFyZW50X2h1Z2VwYWdlcyA9IG1hZHZpc2UK", "source": "/etc/tuned/kernel_settings/tuned.conf", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "src": "/etc/tuned/kernel_settings/tuned.conf"}}} <<< 10371 1726776664.89383: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10371 1726776664.89435: stderr chunk (state=3): >>><<< 10371 1726776664.89443: stdout chunk (state=3): >>><<< 10371 1726776664.89459: _low_level_execute_command() done: rc=0, stdout= {"content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6a2VybmVsX3NldHRpbmdzCgpbbWFpbl0Kc3VtbWFyeSA9IGtlcm5lbCBzZXR0aW5ncwpbc3lzY3RsXQpmcy5lcG9sbC5tYXhfdXNlcl93YXRjaGVzID0gNzg1NTkyCmZzLmZpbGUtbWF4ID0gMzc5NzI0CltzeXNmc10KL3N5cy9rZXJuZWwvZGVidWcveDg2L2licnNfZW5hYmxlZCA9IDAKL3N5cy9rZXJuZWwvZGVidWcveDg2L3B0aV9lbmFibGVkID0gMAovc3lzL2tlcm5lbC9kZWJ1Zy94ODYvcmV0cF9lbmFibGVkID0gMApbdm1dCnRyYW5zcGFyZW50X2h1Z2VwYWdlcyA9IG1hZHZpc2UK", "source": "/etc/tuned/kernel_settings/tuned.conf", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "src": "/etc/tuned/kernel_settings/tuned.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 10371 1726776664.89480: done with _execute_module (slurp, {'path': '/etc/tuned/kernel_settings/tuned.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10371 1726776664.89491: _low_level_execute_command(): starting 10371 1726776664.89496: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776664.6394653-10371-32665271035967/ > /dev/null 2>&1 && sleep 0' 10371 1726776664.92000: stderr chunk (state=2): >>><<< 10371 1726776664.92007: stdout chunk (state=2): >>><<< 10371 1726776664.92021: _low_level_execute_command() done: rc=0, stdout=, stderr= 10371 1726776664.92030: handler run complete 10371 1726776664.92045: attempt loop complete, returning result 10371 1726776664.92050: _execute() done 10371 1726776664.92052: dumping result to json 10371 1726776664.92056: done dumping result, returning 10371 1726776664.92062: done running TaskExecutor() for managed_node3/TASK: Get file [120fa90a-8a95-fee7-0cea-00000000028a] 10371 1726776664.92068: sending task result for task 120fa90a-8a95-fee7-0cea-00000000028a 10371 1726776664.92102: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000028a 10371 1726776664.92106: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6a2VybmVsX3NldHRpbmdzCgpbbWFpbl0Kc3VtbWFyeSA9IGtlcm5lbCBzZXR0aW5ncwpbc3lzY3RsXQpmcy5lcG9sbC5tYXhfdXNlcl93YXRjaGVzID0gNzg1NTkyCmZzLmZpbGUtbWF4ID0gMzc5NzI0CltzeXNmc10KL3N5cy9rZXJuZWwvZGVidWcveDg2L2licnNfZW5hYmxlZCA9IDAKL3N5cy9rZXJuZWwvZGVidWcveDg2L3B0aV9lbmFibGVkID0gMAovc3lzL2tlcm5lbC9kZWJ1Zy94ODYvcmV0cF9lbmFibGVkID0gMApbdm1dCnRyYW5zcGFyZW50X2h1Z2VwYWdlcyA9IG1hZHZpc2UK", "encoding": "base64", "source": "/etc/tuned/kernel_settings/tuned.conf" } 9704 1726776664.92427: no more pending results, returning what we have 9704 1726776664.92432: results queue empty 9704 1726776664.92433: checking for any_errors_fatal 9704 1726776664.92435: done checking for any_errors_fatal 9704 1726776664.92435: checking for max_fail_percentage 9704 1726776664.92437: done checking for max_fail_percentage 9704 1726776664.92438: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.92439: done checking to see if all hosts have failed 9704 1726776664.92439: getting the remaining hosts for this loop 9704 1726776664.92440: done getting the remaining hosts for this loop 9704 1726776664.92444: getting the next task for host managed_node3 9704 1726776664.92449: done getting next task for host managed_node3 9704 1726776664.92451: ^ task is: TASK: Check for presence of ansible managed header, fingerprint 9704 1726776664.92454: ^ state is: HOST STATE: block=2, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.92457: getting variables 9704 1726776664.92458: in VariableManager get_vars() 9704 1726776664.92488: Calling all_inventory to load vars for managed_node3 9704 1726776664.92491: Calling groups_inventory to load vars for managed_node3 9704 1726776664.92493: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.92501: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.92503: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.92506: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.92675: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.92823: done with get_vars() 9704 1726776664.92833: done getting variables 9704 1726776664.92874: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Check for presence of ansible managed header, fingerprint] *************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/check_header.yml:9 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.341) 0:00:17.857 **** 9704 1726776664.92893: entering _queue_task() for managed_node3/assert 9704 1726776664.93060: worker is 1 (out of 1 available) 9704 1726776664.93075: exiting _queue_task() for managed_node3/assert 9704 1726776664.93085: done queuing things up, now waiting for results queue to drain 9704 1726776664.93086: waiting for pending results... 10385 1726776664.93205: running TaskExecutor() for managed_node3/TASK: Check for presence of ansible managed header, fingerprint 10385 1726776664.93303: in run() - task 120fa90a-8a95-fee7-0cea-00000000028b 10385 1726776664.93320: variable 'ansible_search_path' from source: unknown 10385 1726776664.93324: variable 'ansible_search_path' from source: unknown 10385 1726776664.93354: calling self._execute() 10385 1726776664.93418: variable 'ansible_host' from source: host vars for 'managed_node3' 10385 1726776664.93426: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10385 1726776664.93438: variable 'omit' from source: magic vars 10385 1726776664.93512: variable 'omit' from source: magic vars 10385 1726776664.93545: variable 'omit' from source: magic vars 10385 1726776664.93569: variable 'omit' from source: magic vars 10385 1726776664.93599: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10385 1726776664.93626: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10385 1726776664.93650: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10385 1726776664.93666: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10385 1726776664.93677: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10385 1726776664.93700: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10385 1726776664.93705: variable 'ansible_host' from source: host vars for 'managed_node3' 10385 1726776664.93710: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10385 1726776664.93780: Set connection var ansible_module_compression to ZIP_DEFLATED 10385 1726776664.93788: Set connection var ansible_shell_type to sh 10385 1726776664.93797: Set connection var ansible_timeout to 10 10385 1726776664.93803: Set connection var ansible_pipelining to False 10385 1726776664.93809: Set connection var ansible_shell_executable to /bin/sh 10385 1726776664.93812: Set connection var ansible_connection to ssh 10385 1726776664.93827: variable 'ansible_shell_executable' from source: unknown 10385 1726776664.93835: variable 'ansible_connection' from source: unknown 10385 1726776664.93840: variable 'ansible_module_compression' from source: unknown 10385 1726776664.93843: variable 'ansible_shell_type' from source: unknown 10385 1726776664.93847: variable 'ansible_shell_executable' from source: unknown 10385 1726776664.93850: variable 'ansible_host' from source: host vars for 'managed_node3' 10385 1726776664.93854: variable 'ansible_pipelining' from source: unknown 10385 1726776664.93857: variable 'ansible_timeout' from source: unknown 10385 1726776664.93862: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10385 1726776664.93951: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10385 1726776664.93963: variable 'omit' from source: magic vars 10385 1726776664.93969: starting attempt loop 10385 1726776664.93973: running the handler 10385 1726776664.94276: variable 'ansible_managed' from source: task vars 10385 1726776664.94435: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 10385 1726776664.94586: Loaded config def from plugin (lookup/template) 10385 1726776664.94593: Loading LookupModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/template.py 10385 1726776664.94609: File lookup term: get_ansible_managed.j2 10385 1726776664.94613: variable 'ansible_search_path' from source: unknown 10385 1726776664.94618: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings 10385 1726776664.94648: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/templates/get_ansible_managed.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/get_ansible_managed.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/templates/get_ansible_managed.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/get_ansible_managed.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/templates/get_ansible_managed.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/get_ansible_managed.j2 10385 1726776664.94667: variable 'ansible_search_path' from source: unknown 10385 1726776664.95521: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10385 1726776664.97369: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10385 1726776664.97455: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10385 1726776664.97491: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10385 1726776664.97525: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10385 1726776664.97557: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10385 1726776664.97626: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10385 1726776664.97659: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10385 1726776664.97683: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10385 1726776664.97721: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10385 1726776664.97740: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10385 1726776664.97843: variable 'ansible_managed' from source: task vars 10385 1726776664.97899: variable 'content' from source: task vars 10385 1726776664.98052: variable '__content' from source: set_fact 10385 1726776664.98077: Evaluated conditional (ansible_managed in content): True 10385 1726776664.98192: variable '__fingerprint' from source: include params 10385 1726776664.98199: variable 'content' from source: task vars 10385 1726776664.98335: variable '__content' from source: set_fact 10385 1726776664.98351: Evaluated conditional (__fingerprint in content): True 10385 1726776664.98358: handler run complete 10385 1726776664.98371: attempt loop complete, returning result 10385 1726776664.98375: _execute() done 10385 1726776664.98378: dumping result to json 10385 1726776664.98382: done dumping result, returning 10385 1726776664.98388: done running TaskExecutor() for managed_node3/TASK: Check for presence of ansible managed header, fingerprint [120fa90a-8a95-fee7-0cea-00000000028b] 10385 1726776664.98393: sending task result for task 120fa90a-8a95-fee7-0cea-00000000028b 10385 1726776664.98421: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000028b 10385 1726776664.98424: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false } MSG: All assertions passed 9704 1726776664.98849: no more pending results, returning what we have 9704 1726776664.98852: results queue empty 9704 1726776664.98853: checking for any_errors_fatal 9704 1726776664.98860: done checking for any_errors_fatal 9704 1726776664.98861: checking for max_fail_percentage 9704 1726776664.98863: done checking for max_fail_percentage 9704 1726776664.98863: checking to see if all hosts have failed and the running result is not ok 9704 1726776664.98864: done checking to see if all hosts have failed 9704 1726776664.98864: getting the remaining hosts for this loop 9704 1726776664.98866: done getting the remaining hosts for this loop 9704 1726776664.98869: getting the next task for host managed_node3 9704 1726776664.98875: done getting next task for host managed_node3 9704 1726776664.98877: ^ task is: TASK: Ensure role reported changed 9704 1726776664.98879: ^ state is: HOST STATE: block=2, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776664.98882: getting variables 9704 1726776664.98884: in VariableManager get_vars() 9704 1726776664.98914: Calling all_inventory to load vars for managed_node3 9704 1726776664.98916: Calling groups_inventory to load vars for managed_node3 9704 1726776664.98918: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776664.98926: Calling all_plugins_play to load vars for managed_node3 9704 1726776664.98931: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776664.98937: Calling groups_plugins_play to load vars for managed_node3 9704 1726776664.99156: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776664.99352: done with get_vars() 9704 1726776664.99363: done getting variables 9704 1726776664.99418: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Ensure role reported changed] ******************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:41 Thursday 19 September 2024 16:11:04 -0400 (0:00:00.065) 0:00:17.923 **** 9704 1726776664.99447: entering _queue_task() for managed_node3/assert 9704 1726776664.99650: worker is 1 (out of 1 available) 9704 1726776664.99663: exiting _queue_task() for managed_node3/assert 9704 1726776664.99674: done queuing things up, now waiting for results queue to drain 9704 1726776664.99676: waiting for pending results... 10387 1726776664.99882: running TaskExecutor() for managed_node3/TASK: Ensure role reported changed 10387 1726776664.99992: in run() - task 120fa90a-8a95-fee7-0cea-00000000000b 10387 1726776665.00009: variable 'ansible_search_path' from source: unknown 10387 1726776665.00045: calling self._execute() 10387 1726776665.00125: variable 'ansible_host' from source: host vars for 'managed_node3' 10387 1726776665.00138: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10387 1726776665.00148: variable 'omit' from source: magic vars 10387 1726776665.00248: variable 'omit' from source: magic vars 10387 1726776665.00277: variable 'omit' from source: magic vars 10387 1726776665.00307: variable 'omit' from source: magic vars 10387 1726776665.00349: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10387 1726776665.00381: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10387 1726776665.00403: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10387 1726776665.00422: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10387 1726776665.00438: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10387 1726776665.00467: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10387 1726776665.00473: variable 'ansible_host' from source: host vars for 'managed_node3' 10387 1726776665.00477: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10387 1726776665.00577: Set connection var ansible_module_compression to ZIP_DEFLATED 10387 1726776665.00586: Set connection var ansible_shell_type to sh 10387 1726776665.00596: Set connection var ansible_timeout to 10 10387 1726776665.00602: Set connection var ansible_pipelining to False 10387 1726776665.00610: Set connection var ansible_shell_executable to /bin/sh 10387 1726776665.00613: Set connection var ansible_connection to ssh 10387 1726776665.00636: variable 'ansible_shell_executable' from source: unknown 10387 1726776665.00642: variable 'ansible_connection' from source: unknown 10387 1726776665.00646: variable 'ansible_module_compression' from source: unknown 10387 1726776665.00649: variable 'ansible_shell_type' from source: unknown 10387 1726776665.00653: variable 'ansible_shell_executable' from source: unknown 10387 1726776665.00656: variable 'ansible_host' from source: host vars for 'managed_node3' 10387 1726776665.00659: variable 'ansible_pipelining' from source: unknown 10387 1726776665.00662: variable 'ansible_timeout' from source: unknown 10387 1726776665.00666: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10387 1726776665.00869: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10387 1726776665.00882: variable 'omit' from source: magic vars 10387 1726776665.00888: starting attempt loop 10387 1726776665.00892: running the handler 10387 1726776665.01226: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10387 1726776665.03604: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10387 1726776665.03670: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10387 1726776665.03706: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10387 1726776665.03745: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10387 1726776665.03770: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10387 1726776665.03838: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10387 1726776665.03867: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10387 1726776665.03892: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10387 1726776665.03936: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10387 1726776665.03952: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10387 1726776665.04078: variable '__kernel_settings_changed' from source: set_fact 10387 1726776665.04096: Evaluated conditional (__kernel_settings_changed | d(false)): True 10387 1726776665.04103: handler run complete 10387 1726776665.04123: attempt loop complete, returning result 10387 1726776665.04127: _execute() done 10387 1726776665.04136: dumping result to json 10387 1726776665.04142: done dumping result, returning 10387 1726776665.04148: done running TaskExecutor() for managed_node3/TASK: Ensure role reported changed [120fa90a-8a95-fee7-0cea-00000000000b] 10387 1726776665.04154: sending task result for task 120fa90a-8a95-fee7-0cea-00000000000b 10387 1726776665.04182: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000000b 10387 1726776665.04186: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false } MSG: All assertions passed 9704 1726776665.04567: no more pending results, returning what we have 9704 1726776665.04570: results queue empty 9704 1726776665.04571: checking for any_errors_fatal 9704 1726776665.04576: done checking for any_errors_fatal 9704 1726776665.04576: checking for max_fail_percentage 9704 1726776665.04578: done checking for max_fail_percentage 9704 1726776665.04579: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.04579: done checking to see if all hosts have failed 9704 1726776665.04580: getting the remaining hosts for this loop 9704 1726776665.04581: done getting the remaining hosts for this loop 9704 1726776665.04584: getting the next task for host managed_node3 9704 1726776665.04589: done getting next task for host managed_node3 9704 1726776665.04591: ^ task is: TASK: Reboot the machine - see if settings persist after reboot 9704 1726776665.04593: ^ state is: HOST STATE: block=2, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.04596: getting variables 9704 1726776665.04598: in VariableManager get_vars() 9704 1726776665.04627: Calling all_inventory to load vars for managed_node3 9704 1726776665.04632: Calling groups_inventory to load vars for managed_node3 9704 1726776665.04636: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.04644: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.04647: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.04655: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.04825: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.05019: done with get_vars() 9704 1726776665.05031: done getting variables 9704 1726776665.05088: Loading ActionModule 'reboot' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/reboot.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Reboot the machine - see if settings persist after reboot] *************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:47 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.056) 0:00:17.979 **** 9704 1726776665.05117: entering _queue_task() for managed_node3/reboot 9704 1726776665.05292: worker is 1 (out of 1 available) 9704 1726776665.05307: exiting _queue_task() for managed_node3/reboot 9704 1726776665.05318: done queuing things up, now waiting for results queue to drain 9704 1726776665.05320: waiting for pending results... 10389 1726776665.05547: running TaskExecutor() for managed_node3/TASK: Reboot the machine - see if settings persist after reboot 10389 1726776665.05663: in run() - task 120fa90a-8a95-fee7-0cea-00000000000c 10389 1726776665.05682: variable 'ansible_search_path' from source: unknown 10389 1726776665.05716: calling self._execute() 10389 1726776665.05803: variable 'ansible_host' from source: host vars for 'managed_node3' 10389 1726776665.05813: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10389 1726776665.05823: variable 'omit' from source: magic vars 10389 1726776665.06374: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10389 1726776665.08620: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10389 1726776665.08698: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10389 1726776665.08742: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10389 1726776665.08778: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10389 1726776665.08815: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10389 1726776665.08889: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10389 1726776665.08917: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10389 1726776665.08947: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10389 1726776665.08986: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10389 1726776665.09002: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10389 1726776665.09103: variable 'kernel_settings_reboot_required' from source: set_fact 10389 1726776665.09121: Evaluated conditional (kernel_settings_reboot_required | d(false)): False 10389 1726776665.09126: when evaluation is False, skipping this task 10389 1726776665.09132: _execute() done 10389 1726776665.09138: dumping result to json 10389 1726776665.09142: done dumping result, returning 10389 1726776665.09148: done running TaskExecutor() for managed_node3/TASK: Reboot the machine - see if settings persist after reboot [120fa90a-8a95-fee7-0cea-00000000000c] 10389 1726776665.09154: sending task result for task 120fa90a-8a95-fee7-0cea-00000000000c 10389 1726776665.09182: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000000c 10389 1726776665.09186: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "kernel_settings_reboot_required | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776665.09583: no more pending results, returning what we have 9704 1726776665.09586: results queue empty 9704 1726776665.09587: checking for any_errors_fatal 9704 1726776665.09592: done checking for any_errors_fatal 9704 1726776665.09593: checking for max_fail_percentage 9704 1726776665.09595: done checking for max_fail_percentage 9704 1726776665.09595: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.09596: done checking to see if all hosts have failed 9704 1726776665.09596: getting the remaining hosts for this loop 9704 1726776665.09598: done getting the remaining hosts for this loop 9704 1726776665.09601: getting the next task for host managed_node3 9704 1726776665.09606: done getting next task for host managed_node3 9704 1726776665.09609: ^ task is: TASK: Verify that settings were applied correctly after reboot 9704 1726776665.09611: ^ state is: HOST STATE: block=2, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.09613: getting variables 9704 1726776665.09615: in VariableManager get_vars() 9704 1726776665.09650: Calling all_inventory to load vars for managed_node3 9704 1726776665.09653: Calling groups_inventory to load vars for managed_node3 9704 1726776665.09655: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.09664: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.09667: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.09670: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.09898: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.10092: done with get_vars() 9704 1726776665.10102: done getting variables TASK [Verify that settings were applied correctly after reboot] **************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:54 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.050) 0:00:18.030 **** 9704 1726776665.10192: entering _queue_task() for managed_node3/include_tasks 9704 1726776665.10401: worker is 1 (out of 1 available) 9704 1726776665.10415: exiting _queue_task() for managed_node3/include_tasks 9704 1726776665.10426: done queuing things up, now waiting for results queue to drain 9704 1726776665.10428: waiting for pending results... 10390 1726776665.10625: running TaskExecutor() for managed_node3/TASK: Verify that settings were applied correctly after reboot 10390 1726776665.10741: in run() - task 120fa90a-8a95-fee7-0cea-00000000000d 10390 1726776665.10758: variable 'ansible_search_path' from source: unknown 10390 1726776665.10796: calling self._execute() 10390 1726776665.10882: variable 'ansible_host' from source: host vars for 'managed_node3' 10390 1726776665.10892: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10390 1726776665.10900: variable 'omit' from source: magic vars 10390 1726776665.11349: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10390 1726776665.13643: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10390 1726776665.13710: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10390 1726776665.13751: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10390 1726776665.13787: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10390 1726776665.13814: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10390 1726776665.13886: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10390 1726776665.13913: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10390 1726776665.13943: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10390 1726776665.13983: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10390 1726776665.13997: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10390 1726776665.14094: variable 'kernel_settings_reboot_required' from source: set_fact 10390 1726776665.14114: Evaluated conditional (kernel_settings_reboot_required | d(false)): False 10390 1726776665.14118: when evaluation is False, skipping this task 10390 1726776665.14122: _execute() done 10390 1726776665.14125: dumping result to json 10390 1726776665.14130: done dumping result, returning 10390 1726776665.14138: done running TaskExecutor() for managed_node3/TASK: Verify that settings were applied correctly after reboot [120fa90a-8a95-fee7-0cea-00000000000d] 10390 1726776665.14143: sending task result for task 120fa90a-8a95-fee7-0cea-00000000000d 10390 1726776665.14172: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000000d 10390 1726776665.14175: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "kernel_settings_reboot_required | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776665.14526: no more pending results, returning what we have 9704 1726776665.14532: results queue empty 9704 1726776665.14532: checking for any_errors_fatal 9704 1726776665.14540: done checking for any_errors_fatal 9704 1726776665.14541: checking for max_fail_percentage 9704 1726776665.14543: done checking for max_fail_percentage 9704 1726776665.14543: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.14544: done checking to see if all hosts have failed 9704 1726776665.14545: getting the remaining hosts for this loop 9704 1726776665.14546: done getting the remaining hosts for this loop 9704 1726776665.14549: getting the next task for host managed_node3 9704 1726776665.14555: done getting next task for host managed_node3 9704 1726776665.14557: ^ task is: TASK: Apply the settings again to check idempotency 9704 1726776665.14559: ^ state is: HOST STATE: block=2, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.14562: getting variables 9704 1726776665.14563: in VariableManager get_vars() 9704 1726776665.14595: Calling all_inventory to load vars for managed_node3 9704 1726776665.14597: Calling groups_inventory to load vars for managed_node3 9704 1726776665.14599: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.14606: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.14609: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.14611: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.14787: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.14990: done with get_vars() 9704 1726776665.15001: done getting variables TASK [Apply the settings again to check idempotency] *************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:63 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.048) 0:00:18.079 **** 9704 1726776665.15089: entering _queue_task() for managed_node3/include_role 9704 1726776665.15298: worker is 1 (out of 1 available) 9704 1726776665.15313: exiting _queue_task() for managed_node3/include_role 9704 1726776665.15324: done queuing things up, now waiting for results queue to drain 9704 1726776665.15327: waiting for pending results... 10393 1726776665.15540: running TaskExecutor() for managed_node3/TASK: Apply the settings again to check idempotency 10393 1726776665.15656: in run() - task 120fa90a-8a95-fee7-0cea-00000000000e 10393 1726776665.15673: variable 'ansible_search_path' from source: unknown 10393 1726776665.15706: calling self._execute() 10393 1726776665.15795: variable 'ansible_host' from source: host vars for 'managed_node3' 10393 1726776665.15805: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10393 1726776665.15813: variable 'omit' from source: magic vars 10393 1726776665.15920: _execute() done 10393 1726776665.15926: dumping result to json 10393 1726776665.15935: done dumping result, returning 10393 1726776665.15941: done running TaskExecutor() for managed_node3/TASK: Apply the settings again to check idempotency [120fa90a-8a95-fee7-0cea-00000000000e] 10393 1726776665.15949: sending task result for task 120fa90a-8a95-fee7-0cea-00000000000e 10393 1726776665.15980: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000000e 10393 1726776665.15984: WORKER PROCESS EXITING 9704 1726776665.16284: no more pending results, returning what we have 9704 1726776665.16288: in VariableManager get_vars() 9704 1726776665.16320: Calling all_inventory to load vars for managed_node3 9704 1726776665.16323: Calling groups_inventory to load vars for managed_node3 9704 1726776665.16325: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.16340: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.16343: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.16346: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.16560: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.16747: done with get_vars() 9704 1726776665.16753: variable 'ansible_search_path' from source: unknown 9704 1726776665.18228: variable 'omit' from source: magic vars 9704 1726776665.18255: variable 'omit' from source: magic vars 9704 1726776665.18270: variable 'omit' from source: magic vars 9704 1726776665.18273: we have included files to process 9704 1726776665.18274: generating all_blocks data 9704 1726776665.18275: done generating all_blocks data 9704 1726776665.18278: processing included file: fedora.linux_system_roles.kernel_settings 9704 1726776665.18300: in VariableManager get_vars() 9704 1726776665.18315: done with get_vars() 9704 1726776665.18374: in VariableManager get_vars() 9704 1726776665.18391: done with get_vars() 9704 1726776665.18437: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/vars/main.yml 9704 1726776665.18500: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/defaults/main.yml 9704 1726776665.18527: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/meta/main.yml 9704 1726776665.18610: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml 9704 1726776665.19113: in VariableManager get_vars() 9704 1726776665.19138: done with get_vars() 9704 1726776665.20402: in VariableManager get_vars() 9704 1726776665.20424: done with get_vars() 9704 1726776665.20590: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/handlers/main.yml 9704 1726776665.21267: iterating over new_blocks loaded from include file 9704 1726776665.21269: in VariableManager get_vars() 9704 1726776665.21285: done with get_vars() 9704 1726776665.21287: filtering new block on tags 9704 1726776665.21349: done filtering new block on tags 9704 1726776665.21352: in VariableManager get_vars() 9704 1726776665.21367: done with get_vars() 9704 1726776665.21368: filtering new block on tags 9704 1726776665.21407: done filtering new block on tags 9704 1726776665.21409: in VariableManager get_vars() 9704 1726776665.21422: done with get_vars() 9704 1726776665.21424: filtering new block on tags 9704 1726776665.21569: done filtering new block on tags 9704 1726776665.21572: in VariableManager get_vars() 9704 1726776665.21586: done with get_vars() 9704 1726776665.21588: filtering new block on tags 9704 1726776665.21604: done filtering new block on tags 9704 1726776665.21606: done iterating over new_blocks loaded from include file 9704 1726776665.21606: extending task lists for all hosts with included blocks 9704 1726776665.23319: done extending task lists 9704 1726776665.23320: done processing included files 9704 1726776665.23321: results queue empty 9704 1726776665.23321: checking for any_errors_fatal 9704 1726776665.23325: done checking for any_errors_fatal 9704 1726776665.23325: checking for max_fail_percentage 9704 1726776665.23326: done checking for max_fail_percentage 9704 1726776665.23327: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.23327: done checking to see if all hosts have failed 9704 1726776665.23330: getting the remaining hosts for this loop 9704 1726776665.23331: done getting the remaining hosts for this loop 9704 1726776665.23333: getting the next task for host managed_node3 9704 1726776665.23337: done getting next task for host managed_node3 9704 1726776665.23340: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values 9704 1726776665.23342: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.23351: getting variables 9704 1726776665.23352: in VariableManager get_vars() 9704 1726776665.23365: Calling all_inventory to load vars for managed_node3 9704 1726776665.23367: Calling groups_inventory to load vars for managed_node3 9704 1726776665.23369: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.23374: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.23377: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.23379: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.23541: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.23727: done with get_vars() 9704 1726776665.23739: done getting variables 9704 1726776665.23777: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:2 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.087) 0:00:18.166 **** 9704 1726776665.23805: entering _queue_task() for managed_node3/fail 9704 1726776665.24039: worker is 1 (out of 1 available) 9704 1726776665.24053: exiting _queue_task() for managed_node3/fail 9704 1726776665.24064: done queuing things up, now waiting for results queue to drain 9704 1726776665.24067: waiting for pending results... 10396 1726776665.24280: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values 10396 1726776665.24418: in run() - task 120fa90a-8a95-fee7-0cea-0000000003fd 10396 1726776665.24437: variable 'ansible_search_path' from source: unknown 10396 1726776665.24442: variable 'ansible_search_path' from source: unknown 10396 1726776665.24471: calling self._execute() 10396 1726776665.24542: variable 'ansible_host' from source: host vars for 'managed_node3' 10396 1726776665.24551: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10396 1726776665.24560: variable 'omit' from source: magic vars 10396 1726776665.24981: variable 'kernel_settings_sysctl' from source: include_vars 10396 1726776665.25000: variable '__kernel_settings_state_empty' from source: role '' all vars 10396 1726776665.25008: Evaluated conditional (kernel_settings_sysctl != __kernel_settings_state_empty): True 10396 1726776665.25345: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10396 1726776665.27756: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10396 1726776665.27819: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10396 1726776665.27868: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10396 1726776665.27901: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10396 1726776665.27924: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10396 1726776665.27994: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10396 1726776665.28023: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10396 1726776665.28053: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10396 1726776665.28093: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10396 1726776665.28108: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10396 1726776665.28164: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10396 1726776665.28187: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10396 1726776665.28210: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10396 1726776665.28247: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10396 1726776665.28261: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10396 1726776665.28304: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10396 1726776665.28327: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10396 1726776665.28353: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10396 1726776665.28390: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10396 1726776665.28404: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10396 1726776665.28682: variable 'kernel_settings_sysctl' from source: include_vars 10396 1726776665.28757: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10396 1726776665.28969: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10396 1726776665.29008: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10396 1726776665.29040: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10396 1726776665.29069: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10396 1726776665.29112: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10396 1726776665.29136: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10396 1726776665.29161: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10396 1726776665.29186: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10396 1726776665.29230: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10396 1726776665.29252: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10396 1726776665.29275: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10396 1726776665.29300: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10396 1726776665.29327: Evaluated conditional ((kernel_settings_sysctl | selectattr("value", "defined") | selectattr("value", "sameas", true) | list | length > 0) or (kernel_settings_sysctl | selectattr("value", "defined") | selectattr("value", "sameas", false) | list | length > 0)): False 10396 1726776665.29335: when evaluation is False, skipping this task 10396 1726776665.29339: _execute() done 10396 1726776665.29342: dumping result to json 10396 1726776665.29346: done dumping result, returning 10396 1726776665.29352: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values [120fa90a-8a95-fee7-0cea-0000000003fd] 10396 1726776665.29359: sending task result for task 120fa90a-8a95-fee7-0cea-0000000003fd 10396 1726776665.29388: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000003fd 10396 1726776665.29393: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "(kernel_settings_sysctl | selectattr(\"value\", \"defined\") | selectattr(\"value\", \"sameas\", true) | list | length > 0) or (kernel_settings_sysctl | selectattr(\"value\", \"defined\") | selectattr(\"value\", \"sameas\", false) | list | length > 0)", "skip_reason": "Conditional result was False" } 9704 1726776665.29873: no more pending results, returning what we have 9704 1726776665.29876: results queue empty 9704 1726776665.29877: checking for any_errors_fatal 9704 1726776665.29878: done checking for any_errors_fatal 9704 1726776665.29879: checking for max_fail_percentage 9704 1726776665.29881: done checking for max_fail_percentage 9704 1726776665.29881: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.29882: done checking to see if all hosts have failed 9704 1726776665.29882: getting the remaining hosts for this loop 9704 1726776665.29884: done getting the remaining hosts for this loop 9704 1726776665.29887: getting the next task for host managed_node3 9704 1726776665.29893: done getting next task for host managed_node3 9704 1726776665.29896: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables 9704 1726776665.29899: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.29914: getting variables 9704 1726776665.29915: in VariableManager get_vars() 9704 1726776665.29953: Calling all_inventory to load vars for managed_node3 9704 1726776665.29956: Calling groups_inventory to load vars for managed_node3 9704 1726776665.29958: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.29968: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.29970: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.29973: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.30139: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.30347: done with get_vars() 9704 1726776665.30358: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Set version specific variables] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:9 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.066) 0:00:18.233 **** 9704 1726776665.30451: entering _queue_task() for managed_node3/include_tasks 9704 1726776665.30661: worker is 1 (out of 1 available) 9704 1726776665.30673: exiting _queue_task() for managed_node3/include_tasks 9704 1726776665.30684: done queuing things up, now waiting for results queue to drain 9704 1726776665.30686: waiting for pending results... 10399 1726776665.30891: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables 10399 1726776665.31026: in run() - task 120fa90a-8a95-fee7-0cea-0000000003fe 10399 1726776665.31047: variable 'ansible_search_path' from source: unknown 10399 1726776665.31052: variable 'ansible_search_path' from source: unknown 10399 1726776665.31083: calling self._execute() 10399 1726776665.31167: variable 'ansible_host' from source: host vars for 'managed_node3' 10399 1726776665.31176: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10399 1726776665.31185: variable 'omit' from source: magic vars 10399 1726776665.31607: _execute() done 10399 1726776665.31615: dumping result to json 10399 1726776665.31620: done dumping result, returning 10399 1726776665.31626: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables [120fa90a-8a95-fee7-0cea-0000000003fe] 10399 1726776665.31635: sending task result for task 120fa90a-8a95-fee7-0cea-0000000003fe 10399 1726776665.31664: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000003fe 10399 1726776665.31667: WORKER PROCESS EXITING 9704 1726776665.31888: no more pending results, returning what we have 9704 1726776665.31892: in VariableManager get_vars() 9704 1726776665.31936: Calling all_inventory to load vars for managed_node3 9704 1726776665.31940: Calling groups_inventory to load vars for managed_node3 9704 1726776665.31942: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.31953: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.31956: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.31959: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.32375: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.32563: done with get_vars() 9704 1726776665.32570: variable 'ansible_search_path' from source: unknown 9704 1726776665.32571: variable 'ansible_search_path' from source: unknown 9704 1726776665.32607: we have included files to process 9704 1726776665.32609: generating all_blocks data 9704 1726776665.32610: done generating all_blocks data 9704 1726776665.32613: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml 9704 1726776665.32614: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml 9704 1726776665.32616: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml for managed_node3 9704 1726776665.33296: done processing included file 9704 1726776665.33298: iterating over new_blocks loaded from include file 9704 1726776665.33300: in VariableManager get_vars() 9704 1726776665.33322: done with get_vars() 9704 1726776665.33323: filtering new block on tags 9704 1726776665.33352: done filtering new block on tags 9704 1726776665.33354: in VariableManager get_vars() 9704 1726776665.33376: done with get_vars() 9704 1726776665.33378: filtering new block on tags 9704 1726776665.33416: done filtering new block on tags 9704 1726776665.33418: in VariableManager get_vars() 9704 1726776665.33597: done with get_vars() 9704 1726776665.33599: filtering new block on tags 9704 1726776665.33665: done filtering new block on tags 9704 1726776665.33668: in VariableManager get_vars() 9704 1726776665.33688: done with get_vars() 9704 1726776665.33690: filtering new block on tags 9704 1726776665.33712: done filtering new block on tags 9704 1726776665.33714: done iterating over new_blocks loaded from include file 9704 1726776665.33715: extending task lists for all hosts with included blocks 9704 1726776665.33868: done extending task lists 9704 1726776665.33870: done processing included files 9704 1726776665.33870: results queue empty 9704 1726776665.33871: checking for any_errors_fatal 9704 1726776665.33874: done checking for any_errors_fatal 9704 1726776665.33875: checking for max_fail_percentage 9704 1726776665.33876: done checking for max_fail_percentage 9704 1726776665.33876: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.33877: done checking to see if all hosts have failed 9704 1726776665.33878: getting the remaining hosts for this loop 9704 1726776665.33878: done getting the remaining hosts for this loop 9704 1726776665.33881: getting the next task for host managed_node3 9704 1726776665.33885: done getting next task for host managed_node3 9704 1726776665.33887: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role 9704 1726776665.33890: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.33899: getting variables 9704 1726776665.33900: in VariableManager get_vars() 9704 1726776665.33912: Calling all_inventory to load vars for managed_node3 9704 1726776665.33914: Calling groups_inventory to load vars for managed_node3 9704 1726776665.33916: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.33921: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.33923: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.33926: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.34055: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.34233: done with get_vars() 9704 1726776665.34241: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:2 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.038) 0:00:18.271 **** 9704 1726776665.34301: entering _queue_task() for managed_node3/setup 9704 1726776665.34516: worker is 1 (out of 1 available) 9704 1726776665.34531: exiting _queue_task() for managed_node3/setup 9704 1726776665.34543: done queuing things up, now waiting for results queue to drain 9704 1726776665.34545: waiting for pending results... 10401 1726776665.34761: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role 10401 1726776665.34913: in run() - task 120fa90a-8a95-fee7-0cea-00000000050b 10401 1726776665.34933: variable 'ansible_search_path' from source: unknown 10401 1726776665.34938: variable 'ansible_search_path' from source: unknown 10401 1726776665.34968: calling self._execute() 10401 1726776665.35044: variable 'ansible_host' from source: host vars for 'managed_node3' 10401 1726776665.35053: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10401 1726776665.35061: variable 'omit' from source: magic vars 10401 1726776665.35614: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10401 1726776665.37826: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10401 1726776665.37898: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10401 1726776665.37932: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10401 1726776665.37976: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10401 1726776665.37999: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10401 1726776665.38066: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10401 1726776665.38093: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10401 1726776665.38118: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10401 1726776665.38154: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10401 1726776665.38166: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10401 1726776665.38214: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10401 1726776665.38236: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10401 1726776665.38256: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10401 1726776665.38288: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10401 1726776665.38298: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10401 1726776665.38454: variable '__kernel_settings_required_facts' from source: role '' all vars 10401 1726776665.38465: variable 'ansible_facts' from source: unknown 10401 1726776665.38546: Evaluated conditional (__kernel_settings_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 10401 1726776665.38552: when evaluation is False, skipping this task 10401 1726776665.38556: _execute() done 10401 1726776665.38558: dumping result to json 10401 1726776665.38562: done dumping result, returning 10401 1726776665.38568: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role [120fa90a-8a95-fee7-0cea-00000000050b] 10401 1726776665.38574: sending task result for task 120fa90a-8a95-fee7-0cea-00000000050b 10401 1726776665.38605: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000050b 10401 1726776665.38609: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } 9704 1726776665.39058: no more pending results, returning what we have 9704 1726776665.39061: results queue empty 9704 1726776665.39062: checking for any_errors_fatal 9704 1726776665.39064: done checking for any_errors_fatal 9704 1726776665.39065: checking for max_fail_percentage 9704 1726776665.39066: done checking for max_fail_percentage 9704 1726776665.39067: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.39068: done checking to see if all hosts have failed 9704 1726776665.39068: getting the remaining hosts for this loop 9704 1726776665.39069: done getting the remaining hosts for this loop 9704 1726776665.39073: getting the next task for host managed_node3 9704 1726776665.39082: done getting next task for host managed_node3 9704 1726776665.39086: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree 9704 1726776665.39089: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.39104: getting variables 9704 1726776665.39106: in VariableManager get_vars() 9704 1726776665.39146: Calling all_inventory to load vars for managed_node3 9704 1726776665.39149: Calling groups_inventory to load vars for managed_node3 9704 1726776665.39151: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.39161: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.39163: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.39166: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.39384: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.39600: done with get_vars() 9704 1726776665.39612: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Check if system is ostree] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:10 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.054) 0:00:18.325 **** 9704 1726776665.39712: entering _queue_task() for managed_node3/stat 9704 1726776665.39977: worker is 1 (out of 1 available) 9704 1726776665.39991: exiting _queue_task() for managed_node3/stat 9704 1726776665.40003: done queuing things up, now waiting for results queue to drain 9704 1726776665.40006: waiting for pending results... 10403 1726776665.40250: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree 10403 1726776665.40409: in run() - task 120fa90a-8a95-fee7-0cea-00000000050d 10403 1726776665.40431: variable 'ansible_search_path' from source: unknown 10403 1726776665.40437: variable 'ansible_search_path' from source: unknown 10403 1726776665.40470: calling self._execute() 10403 1726776665.40554: variable 'ansible_host' from source: host vars for 'managed_node3' 10403 1726776665.40563: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10403 1726776665.40572: variable 'omit' from source: magic vars 10403 1726776665.41007: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10403 1726776665.41271: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10403 1726776665.41304: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10403 1726776665.41336: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10403 1726776665.41368: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10403 1726776665.41428: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10403 1726776665.41452: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10403 1726776665.41473: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10403 1726776665.41491: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10403 1726776665.41578: variable '__kernel_settings_is_ostree' from source: set_fact 10403 1726776665.41589: Evaluated conditional (not __kernel_settings_is_ostree is defined): False 10403 1726776665.41594: when evaluation is False, skipping this task 10403 1726776665.41597: _execute() done 10403 1726776665.41601: dumping result to json 10403 1726776665.41605: done dumping result, returning 10403 1726776665.41611: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree [120fa90a-8a95-fee7-0cea-00000000050d] 10403 1726776665.41616: sending task result for task 120fa90a-8a95-fee7-0cea-00000000050d 10403 1726776665.41642: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000050d 10403 1726776665.41646: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_ostree is defined", "skip_reason": "Conditional result was False" } 9704 1726776665.41751: no more pending results, returning what we have 9704 1726776665.41753: results queue empty 9704 1726776665.41754: checking for any_errors_fatal 9704 1726776665.41760: done checking for any_errors_fatal 9704 1726776665.41760: checking for max_fail_percentage 9704 1726776665.41762: done checking for max_fail_percentage 9704 1726776665.41762: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.41763: done checking to see if all hosts have failed 9704 1726776665.41763: getting the remaining hosts for this loop 9704 1726776665.41765: done getting the remaining hosts for this loop 9704 1726776665.41767: getting the next task for host managed_node3 9704 1726776665.41773: done getting next task for host managed_node3 9704 1726776665.41776: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree 9704 1726776665.41779: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.41793: getting variables 9704 1726776665.41794: in VariableManager get_vars() 9704 1726776665.41825: Calling all_inventory to load vars for managed_node3 9704 1726776665.41830: Calling groups_inventory to load vars for managed_node3 9704 1726776665.41832: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.41840: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.41842: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.41845: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.41958: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.42077: done with get_vars() 9704 1726776665.42085: done getting variables 9704 1726776665.42123: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:15 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.024) 0:00:18.350 **** 9704 1726776665.42147: entering _queue_task() for managed_node3/set_fact 9704 1726776665.42300: worker is 1 (out of 1 available) 9704 1726776665.42313: exiting _queue_task() for managed_node3/set_fact 9704 1726776665.42325: done queuing things up, now waiting for results queue to drain 9704 1726776665.42327: waiting for pending results... 10405 1726776665.42447: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree 10405 1726776665.42563: in run() - task 120fa90a-8a95-fee7-0cea-00000000050e 10405 1726776665.42580: variable 'ansible_search_path' from source: unknown 10405 1726776665.42584: variable 'ansible_search_path' from source: unknown 10405 1726776665.42610: calling self._execute() 10405 1726776665.42672: variable 'ansible_host' from source: host vars for 'managed_node3' 10405 1726776665.42680: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10405 1726776665.42689: variable 'omit' from source: magic vars 10405 1726776665.43000: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10405 1726776665.43295: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10405 1726776665.43325: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10405 1726776665.43365: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10405 1726776665.43392: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10405 1726776665.43462: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10405 1726776665.43487: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10405 1726776665.43513: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10405 1726776665.43543: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10405 1726776665.43653: variable '__kernel_settings_is_ostree' from source: set_fact 10405 1726776665.43665: Evaluated conditional (not __kernel_settings_is_ostree is defined): False 10405 1726776665.43669: when evaluation is False, skipping this task 10405 1726776665.43672: _execute() done 10405 1726776665.43675: dumping result to json 10405 1726776665.43678: done dumping result, returning 10405 1726776665.43683: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree [120fa90a-8a95-fee7-0cea-00000000050e] 10405 1726776665.43688: sending task result for task 120fa90a-8a95-fee7-0cea-00000000050e 10405 1726776665.43713: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000050e 10405 1726776665.43716: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_ostree is defined", "skip_reason": "Conditional result was False" } 9704 1726776665.44035: no more pending results, returning what we have 9704 1726776665.44038: results queue empty 9704 1726776665.44038: checking for any_errors_fatal 9704 1726776665.44042: done checking for any_errors_fatal 9704 1726776665.44042: checking for max_fail_percentage 9704 1726776665.44043: done checking for max_fail_percentage 9704 1726776665.44044: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.44044: done checking to see if all hosts have failed 9704 1726776665.44044: getting the remaining hosts for this loop 9704 1726776665.44046: done getting the remaining hosts for this loop 9704 1726776665.44048: getting the next task for host managed_node3 9704 1726776665.44054: done getting next task for host managed_node3 9704 1726776665.44056: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin 9704 1726776665.44059: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.44069: getting variables 9704 1726776665.44070: in VariableManager get_vars() 9704 1726776665.44093: Calling all_inventory to load vars for managed_node3 9704 1726776665.44095: Calling groups_inventory to load vars for managed_node3 9704 1726776665.44097: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.44103: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.44105: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.44106: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.44253: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.44372: done with get_vars() 9704 1726776665.44379: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:22 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.022) 0:00:18.373 **** 9704 1726776665.44442: entering _queue_task() for managed_node3/stat 9704 1726776665.44598: worker is 1 (out of 1 available) 9704 1726776665.44613: exiting _queue_task() for managed_node3/stat 9704 1726776665.44624: done queuing things up, now waiting for results queue to drain 9704 1726776665.44626: waiting for pending results... 10407 1726776665.44743: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin 10407 1726776665.44862: in run() - task 120fa90a-8a95-fee7-0cea-000000000510 10407 1726776665.44878: variable 'ansible_search_path' from source: unknown 10407 1726776665.44882: variable 'ansible_search_path' from source: unknown 10407 1726776665.44908: calling self._execute() 10407 1726776665.44974: variable 'ansible_host' from source: host vars for 'managed_node3' 10407 1726776665.44982: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10407 1726776665.44991: variable 'omit' from source: magic vars 10407 1726776665.45308: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10407 1726776665.45488: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10407 1726776665.45523: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10407 1726776665.45552: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10407 1726776665.45580: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10407 1726776665.45644: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10407 1726776665.45664: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10407 1726776665.45683: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10407 1726776665.45700: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10407 1726776665.45787: variable '__kernel_settings_is_transactional' from source: set_fact 10407 1726776665.45799: Evaluated conditional (not __kernel_settings_is_transactional is defined): False 10407 1726776665.45803: when evaluation is False, skipping this task 10407 1726776665.45806: _execute() done 10407 1726776665.45810: dumping result to json 10407 1726776665.45813: done dumping result, returning 10407 1726776665.45820: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin [120fa90a-8a95-fee7-0cea-000000000510] 10407 1726776665.45825: sending task result for task 120fa90a-8a95-fee7-0cea-000000000510 10407 1726776665.45852: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000510 10407 1726776665.45856: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_transactional is defined", "skip_reason": "Conditional result was False" } 9704 1726776665.45977: no more pending results, returning what we have 9704 1726776665.45980: results queue empty 9704 1726776665.45981: checking for any_errors_fatal 9704 1726776665.45986: done checking for any_errors_fatal 9704 1726776665.45987: checking for max_fail_percentage 9704 1726776665.45988: done checking for max_fail_percentage 9704 1726776665.45989: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.45989: done checking to see if all hosts have failed 9704 1726776665.45990: getting the remaining hosts for this loop 9704 1726776665.45991: done getting the remaining hosts for this loop 9704 1726776665.45994: getting the next task for host managed_node3 9704 1726776665.46000: done getting next task for host managed_node3 9704 1726776665.46003: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists 9704 1726776665.46006: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.46020: getting variables 9704 1726776665.46021: in VariableManager get_vars() 9704 1726776665.46056: Calling all_inventory to load vars for managed_node3 9704 1726776665.46058: Calling groups_inventory to load vars for managed_node3 9704 1726776665.46060: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.46066: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.46067: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.46069: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.46177: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.46352: done with get_vars() 9704 1726776665.46362: done getting variables 9704 1726776665.46413: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:27 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.020) 0:00:18.393 **** 9704 1726776665.46449: entering _queue_task() for managed_node3/set_fact 9704 1726776665.46657: worker is 1 (out of 1 available) 9704 1726776665.46670: exiting _queue_task() for managed_node3/set_fact 9704 1726776665.46682: done queuing things up, now waiting for results queue to drain 9704 1726776665.46684: waiting for pending results... 10408 1726776665.46890: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists 10408 1726776665.47046: in run() - task 120fa90a-8a95-fee7-0cea-000000000511 10408 1726776665.47065: variable 'ansible_search_path' from source: unknown 10408 1726776665.47069: variable 'ansible_search_path' from source: unknown 10408 1726776665.47101: calling self._execute() 10408 1726776665.47179: variable 'ansible_host' from source: host vars for 'managed_node3' 10408 1726776665.47188: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10408 1726776665.47196: variable 'omit' from source: magic vars 10408 1726776665.47539: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10408 1726776665.47836: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10408 1726776665.47875: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10408 1726776665.47908: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10408 1726776665.47943: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10408 1726776665.48014: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10408 1726776665.48045: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10408 1726776665.48071: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10408 1726776665.48097: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10408 1726776665.48210: variable '__kernel_settings_is_transactional' from source: set_fact 10408 1726776665.48222: Evaluated conditional (not __kernel_settings_is_transactional is defined): False 10408 1726776665.48226: when evaluation is False, skipping this task 10408 1726776665.48231: _execute() done 10408 1726776665.48237: dumping result to json 10408 1726776665.48240: done dumping result, returning 10408 1726776665.48246: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists [120fa90a-8a95-fee7-0cea-000000000511] 10408 1726776665.48251: sending task result for task 120fa90a-8a95-fee7-0cea-000000000511 10408 1726776665.48272: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000511 10408 1726776665.48276: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_transactional is defined", "skip_reason": "Conditional result was False" } 9704 1726776665.48376: no more pending results, returning what we have 9704 1726776665.48378: results queue empty 9704 1726776665.48379: checking for any_errors_fatal 9704 1726776665.48384: done checking for any_errors_fatal 9704 1726776665.48385: checking for max_fail_percentage 9704 1726776665.48386: done checking for max_fail_percentage 9704 1726776665.48386: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.48387: done checking to see if all hosts have failed 9704 1726776665.48388: getting the remaining hosts for this loop 9704 1726776665.48389: done getting the remaining hosts for this loop 9704 1726776665.48392: getting the next task for host managed_node3 9704 1726776665.48400: done getting next task for host managed_node3 9704 1726776665.48403: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables 9704 1726776665.48406: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.48420: getting variables 9704 1726776665.48421: in VariableManager get_vars() 9704 1726776665.48452: Calling all_inventory to load vars for managed_node3 9704 1726776665.48454: Calling groups_inventory to load vars for managed_node3 9704 1726776665.48456: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.48463: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.48465: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.48468: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.48609: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.48724: done with get_vars() 9704 1726776665.48734: done getting variables 9704 1726776665.48776: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set platform/version specific variables] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:31 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.023) 0:00:18.416 **** 9704 1726776665.48798: entering _queue_task() for managed_node3/include_vars 9704 1726776665.48947: worker is 1 (out of 1 available) 9704 1726776665.48961: exiting _queue_task() for managed_node3/include_vars 9704 1726776665.48973: done queuing things up, now waiting for results queue to drain 9704 1726776665.48975: waiting for pending results... 10411 1726776665.49090: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables 10411 1726776665.49211: in run() - task 120fa90a-8a95-fee7-0cea-000000000513 10411 1726776665.49226: variable 'ansible_search_path' from source: unknown 10411 1726776665.49232: variable 'ansible_search_path' from source: unknown 10411 1726776665.49259: calling self._execute() 10411 1726776665.49320: variable 'ansible_host' from source: host vars for 'managed_node3' 10411 1726776665.49326: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10411 1726776665.49333: variable 'omit' from source: magic vars 10411 1726776665.49402: variable 'omit' from source: magic vars 10411 1726776665.49456: variable 'omit' from source: magic vars 10411 1726776665.49706: variable 'ffparams' from source: task vars 10411 1726776665.49806: variable 'ansible_facts' from source: unknown 10411 1726776665.49940: variable 'ansible_facts' from source: unknown 10411 1726776665.50026: variable 'ansible_facts' from source: unknown 10411 1726776665.50116: variable 'ansible_facts' from source: unknown 10411 1726776665.50196: variable 'role_path' from source: magic vars 10411 1726776665.50312: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 10411 1726776665.50470: Loaded config def from plugin (lookup/first_found) 10411 1726776665.50478: Loading LookupModule 'first_found' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/first_found.py 10411 1726776665.50505: variable 'ansible_search_path' from source: unknown 10411 1726776665.50523: variable 'ansible_search_path' from source: unknown 10411 1726776665.50533: variable 'ansible_search_path' from source: unknown 10411 1726776665.50543: variable 'ansible_search_path' from source: unknown 10411 1726776665.50551: variable 'ansible_search_path' from source: unknown 10411 1726776665.50565: variable 'omit' from source: magic vars 10411 1726776665.50582: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10411 1726776665.50603: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10411 1726776665.50620: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10411 1726776665.50638: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10411 1726776665.50652: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10411 1726776665.50674: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10411 1726776665.50679: variable 'ansible_host' from source: host vars for 'managed_node3' 10411 1726776665.50683: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10411 1726776665.50770: Set connection var ansible_module_compression to ZIP_DEFLATED 10411 1726776665.50776: Set connection var ansible_shell_type to sh 10411 1726776665.50781: Set connection var ansible_timeout to 10 10411 1726776665.50787: Set connection var ansible_pipelining to False 10411 1726776665.50795: Set connection var ansible_shell_executable to /bin/sh 10411 1726776665.50798: Set connection var ansible_connection to ssh 10411 1726776665.50815: variable 'ansible_shell_executable' from source: unknown 10411 1726776665.50817: variable 'ansible_connection' from source: unknown 10411 1726776665.50819: variable 'ansible_module_compression' from source: unknown 10411 1726776665.50821: variable 'ansible_shell_type' from source: unknown 10411 1726776665.50822: variable 'ansible_shell_executable' from source: unknown 10411 1726776665.50824: variable 'ansible_host' from source: host vars for 'managed_node3' 10411 1726776665.50826: variable 'ansible_pipelining' from source: unknown 10411 1726776665.50828: variable 'ansible_timeout' from source: unknown 10411 1726776665.50833: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10411 1726776665.50924: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10411 1726776665.50940: variable 'omit' from source: magic vars 10411 1726776665.50946: starting attempt loop 10411 1726776665.50949: running the handler 10411 1726776665.51003: handler run complete 10411 1726776665.51013: attempt loop complete, returning result 10411 1726776665.51017: _execute() done 10411 1726776665.51020: dumping result to json 10411 1726776665.51024: done dumping result, returning 10411 1726776665.51052: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables [120fa90a-8a95-fee7-0cea-000000000513] 10411 1726776665.51060: sending task result for task 120fa90a-8a95-fee7-0cea-000000000513 10411 1726776665.51088: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000513 10411 1726776665.51091: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_packages": [ "tuned", "python3-configobj" ], "__kernel_settings_services": [ "tuned" ] }, "ansible_included_var_files": [ "/tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/vars/default.yml" ], "changed": false } 9704 1726776665.51519: no more pending results, returning what we have 9704 1726776665.51521: results queue empty 9704 1726776665.51524: checking for any_errors_fatal 9704 1726776665.51530: done checking for any_errors_fatal 9704 1726776665.51531: checking for max_fail_percentage 9704 1726776665.51532: done checking for max_fail_percentage 9704 1726776665.51533: checking to see if all hosts have failed and the running result is not ok 9704 1726776665.51535: done checking to see if all hosts have failed 9704 1726776665.51535: getting the remaining hosts for this loop 9704 1726776665.51537: done getting the remaining hosts for this loop 9704 1726776665.51539: getting the next task for host managed_node3 9704 1726776665.51549: done getting next task for host managed_node3 9704 1726776665.51553: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed 9704 1726776665.51555: ^ state is: HOST STATE: block=2, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776665.51565: getting variables 9704 1726776665.51566: in VariableManager get_vars() 9704 1726776665.51594: Calling all_inventory to load vars for managed_node3 9704 1726776665.51596: Calling groups_inventory to load vars for managed_node3 9704 1726776665.51598: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776665.51609: Calling all_plugins_play to load vars for managed_node3 9704 1726776665.51613: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776665.51616: Calling groups_plugins_play to load vars for managed_node3 9704 1726776665.51790: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776665.52005: done with get_vars() 9704 1726776665.52014: done getting variables 9704 1726776665.52072: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure required packages are installed] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:12 Thursday 19 September 2024 16:11:05 -0400 (0:00:00.033) 0:00:18.449 **** 9704 1726776665.52103: entering _queue_task() for managed_node3/package 9704 1726776665.52274: worker is 1 (out of 1 available) 9704 1726776665.52289: exiting _queue_task() for managed_node3/package 9704 1726776665.52302: done queuing things up, now waiting for results queue to drain 9704 1726776665.52304: waiting for pending results... 10412 1726776665.52504: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed 10412 1726776665.52627: in run() - task 120fa90a-8a95-fee7-0cea-0000000003ff 10412 1726776665.52653: variable 'ansible_search_path' from source: unknown 10412 1726776665.52657: variable 'ansible_search_path' from source: unknown 10412 1726776665.52689: calling self._execute() 10412 1726776665.52774: variable 'ansible_host' from source: host vars for 'managed_node3' 10412 1726776665.52785: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10412 1726776665.52793: variable 'omit' from source: magic vars 10412 1726776665.52887: variable 'omit' from source: magic vars 10412 1726776665.52937: variable 'omit' from source: magic vars 10412 1726776665.52964: variable '__kernel_settings_packages' from source: include_vars 10412 1726776665.53294: variable '__kernel_settings_packages' from source: include_vars 10412 1726776665.53462: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10412 1726776665.54904: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10412 1726776665.54959: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10412 1726776665.54988: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10412 1726776665.55016: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10412 1726776665.55038: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10412 1726776665.55105: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10412 1726776665.55126: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10412 1726776665.55148: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10412 1726776665.55175: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10412 1726776665.55188: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10412 1726776665.55289: variable '__kernel_settings_is_ostree' from source: set_fact 10412 1726776665.55296: variable 'omit' from source: magic vars 10412 1726776665.55316: variable 'omit' from source: magic vars 10412 1726776665.55335: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10412 1726776665.55353: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10412 1726776665.55365: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10412 1726776665.55375: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10412 1726776665.55382: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10412 1726776665.55404: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10412 1726776665.55408: variable 'ansible_host' from source: host vars for 'managed_node3' 10412 1726776665.55410: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10412 1726776665.55491: Set connection var ansible_module_compression to ZIP_DEFLATED 10412 1726776665.55499: Set connection var ansible_shell_type to sh 10412 1726776665.55509: Set connection var ansible_timeout to 10 10412 1726776665.55515: Set connection var ansible_pipelining to False 10412 1726776665.55522: Set connection var ansible_shell_executable to /bin/sh 10412 1726776665.55525: Set connection var ansible_connection to ssh 10412 1726776665.55563: variable 'ansible_shell_executable' from source: unknown 10412 1726776665.55569: variable 'ansible_connection' from source: unknown 10412 1726776665.55572: variable 'ansible_module_compression' from source: unknown 10412 1726776665.55576: variable 'ansible_shell_type' from source: unknown 10412 1726776665.55579: variable 'ansible_shell_executable' from source: unknown 10412 1726776665.55582: variable 'ansible_host' from source: host vars for 'managed_node3' 10412 1726776665.55586: variable 'ansible_pipelining' from source: unknown 10412 1726776665.55590: variable 'ansible_timeout' from source: unknown 10412 1726776665.55594: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10412 1726776665.55670: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10412 1726776665.55683: variable 'omit' from source: magic vars 10412 1726776665.55690: starting attempt loop 10412 1726776665.55694: running the handler 10412 1726776665.55776: variable 'ansible_facts' from source: unknown 10412 1726776665.55885: _low_level_execute_command(): starting 10412 1726776665.55894: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10412 1726776665.58258: stdout chunk (state=2): >>>/root <<< 10412 1726776665.58382: stderr chunk (state=3): >>><<< 10412 1726776665.58389: stdout chunk (state=3): >>><<< 10412 1726776665.58408: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10412 1726776665.58420: _low_level_execute_command(): starting 10412 1726776665.58426: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160 `" && echo ansible-tmp-1726776665.5841615-10412-273802752060160="` echo /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160 `" ) && sleep 0' 10412 1726776665.60938: stdout chunk (state=2): >>>ansible-tmp-1726776665.5841615-10412-273802752060160=/root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160 <<< 10412 1726776665.61067: stderr chunk (state=3): >>><<< 10412 1726776665.61073: stdout chunk (state=3): >>><<< 10412 1726776665.61088: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776665.5841615-10412-273802752060160=/root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160 , stderr= 10412 1726776665.61112: variable 'ansible_module_compression' from source: unknown 10412 1726776665.61155: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 10412 1726776665.61191: variable 'ansible_facts' from source: unknown 10412 1726776665.61280: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/AnsiballZ_dnf.py 10412 1726776665.61375: Sending initial data 10412 1726776665.61382: Sent initial data (151 bytes) 10412 1726776665.63909: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpmj4cmn0o /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/AnsiballZ_dnf.py <<< 10412 1726776665.65161: stderr chunk (state=3): >>><<< 10412 1726776665.65167: stdout chunk (state=3): >>><<< 10412 1726776665.65186: done transferring module to remote 10412 1726776665.65197: _low_level_execute_command(): starting 10412 1726776665.65202: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/ /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/AnsiballZ_dnf.py && sleep 0' 10412 1726776665.67556: stderr chunk (state=2): >>><<< 10412 1726776665.67563: stdout chunk (state=2): >>><<< 10412 1726776665.67576: _low_level_execute_command() done: rc=0, stdout=, stderr= 10412 1726776665.67580: _low_level_execute_command(): starting 10412 1726776665.67585: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/AnsiballZ_dnf.py && sleep 0' 10412 1726776668.25498: stdout chunk (state=2): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["tuned", "python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} <<< 10412 1726776668.34025: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10412 1726776668.34079: stderr chunk (state=3): >>><<< 10412 1726776668.34087: stdout chunk (state=3): >>><<< 10412 1726776668.34103: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["tuned", "python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10412 1726776668.34139: done with _execute_module (ansible.legacy.dnf, {'name': ['tuned', 'python3-configobj'], 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10412 1726776668.34149: _low_level_execute_command(): starting 10412 1726776668.34155: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776665.5841615-10412-273802752060160/ > /dev/null 2>&1 && sleep 0' 10412 1726776668.36586: stderr chunk (state=2): >>><<< 10412 1726776668.36594: stdout chunk (state=2): >>><<< 10412 1726776668.36608: _low_level_execute_command() done: rc=0, stdout=, stderr= 10412 1726776668.36615: handler run complete 10412 1726776668.36645: attempt loop complete, returning result 10412 1726776668.36650: _execute() done 10412 1726776668.36653: dumping result to json 10412 1726776668.36659: done dumping result, returning 10412 1726776668.36666: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed [120fa90a-8a95-fee7-0cea-0000000003ff] 10412 1726776668.36671: sending task result for task 120fa90a-8a95-fee7-0cea-0000000003ff 10412 1726776668.36703: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000003ff 10412 1726776668.36706: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 9704 1726776668.36864: no more pending results, returning what we have 9704 1726776668.36867: results queue empty 9704 1726776668.36868: checking for any_errors_fatal 9704 1726776668.36875: done checking for any_errors_fatal 9704 1726776668.36875: checking for max_fail_percentage 9704 1726776668.36876: done checking for max_fail_percentage 9704 1726776668.36877: checking to see if all hosts have failed and the running result is not ok 9704 1726776668.36877: done checking to see if all hosts have failed 9704 1726776668.36878: getting the remaining hosts for this loop 9704 1726776668.36879: done getting the remaining hosts for this loop 9704 1726776668.36882: getting the next task for host managed_node3 9704 1726776668.36890: done getting next task for host managed_node3 9704 1726776668.36893: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes 9704 1726776668.36895: ^ state is: HOST STATE: block=2, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776668.36905: getting variables 9704 1726776668.36906: in VariableManager get_vars() 9704 1726776668.36940: Calling all_inventory to load vars for managed_node3 9704 1726776668.36943: Calling groups_inventory to load vars for managed_node3 9704 1726776668.36945: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776668.36953: Calling all_plugins_play to load vars for managed_node3 9704 1726776668.36955: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776668.36958: Calling groups_plugins_play to load vars for managed_node3 9704 1726776668.37114: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776668.37235: done with get_vars() 9704 1726776668.37243: done getting variables 9704 1726776668.37283: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:24 Thursday 19 September 2024 16:11:08 -0400 (0:00:02.852) 0:00:21.301 **** 9704 1726776668.37304: entering _queue_task() for managed_node3/debug 9704 1726776668.37472: worker is 1 (out of 1 available) 9704 1726776668.37485: exiting _queue_task() for managed_node3/debug 9704 1726776668.37497: done queuing things up, now waiting for results queue to drain 9704 1726776668.37499: waiting for pending results... 10431 1726776668.37619: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes 10431 1726776668.37736: in run() - task 120fa90a-8a95-fee7-0cea-000000000401 10431 1726776668.37754: variable 'ansible_search_path' from source: unknown 10431 1726776668.37758: variable 'ansible_search_path' from source: unknown 10431 1726776668.37785: calling self._execute() 10431 1726776668.37856: variable 'ansible_host' from source: host vars for 'managed_node3' 10431 1726776668.37863: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10431 1726776668.37869: variable 'omit' from source: magic vars 10431 1726776668.38201: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10431 1726776668.39713: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10431 1726776668.39762: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10431 1726776668.39792: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10431 1726776668.39820: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10431 1726776668.39844: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10431 1726776668.39898: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10431 1726776668.39919: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10431 1726776668.39942: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10431 1726776668.39971: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10431 1726776668.39982: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10431 1726776668.40060: variable '__kernel_settings_is_transactional' from source: set_fact 10431 1726776668.40076: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 10431 1726776668.40080: when evaluation is False, skipping this task 10431 1726776668.40083: _execute() done 10431 1726776668.40087: dumping result to json 10431 1726776668.40091: done dumping result, returning 10431 1726776668.40097: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes [120fa90a-8a95-fee7-0cea-000000000401] 10431 1726776668.40102: sending task result for task 120fa90a-8a95-fee7-0cea-000000000401 10431 1726776668.40122: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000401 10431 1726776668.40124: WORKER PROCESS EXITING skipping: [managed_node3] => { "false_condition": "__kernel_settings_is_transactional | d(false)" } 9704 1726776668.40349: no more pending results, returning what we have 9704 1726776668.40351: results queue empty 9704 1726776668.40352: checking for any_errors_fatal 9704 1726776668.40358: done checking for any_errors_fatal 9704 1726776668.40358: checking for max_fail_percentage 9704 1726776668.40359: done checking for max_fail_percentage 9704 1726776668.40359: checking to see if all hosts have failed and the running result is not ok 9704 1726776668.40360: done checking to see if all hosts have failed 9704 1726776668.40360: getting the remaining hosts for this loop 9704 1726776668.40361: done getting the remaining hosts for this loop 9704 1726776668.40364: getting the next task for host managed_node3 9704 1726776668.40369: done getting next task for host managed_node3 9704 1726776668.40372: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems 9704 1726776668.40373: ^ state is: HOST STATE: block=2, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776668.40383: getting variables 9704 1726776668.40384: in VariableManager get_vars() 9704 1726776668.40407: Calling all_inventory to load vars for managed_node3 9704 1726776668.40408: Calling groups_inventory to load vars for managed_node3 9704 1726776668.40409: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776668.40415: Calling all_plugins_play to load vars for managed_node3 9704 1726776668.40417: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776668.40418: Calling groups_plugins_play to load vars for managed_node3 9704 1726776668.40522: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776668.40647: done with get_vars() 9704 1726776668.40655: done getting variables 9704 1726776668.40693: Loading ActionModule 'reboot' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/reboot.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Reboot transactional update systems] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:29 Thursday 19 September 2024 16:11:08 -0400 (0:00:00.034) 0:00:21.335 **** 9704 1726776668.40713: entering _queue_task() for managed_node3/reboot 9704 1726776668.40861: worker is 1 (out of 1 available) 9704 1726776668.40874: exiting _queue_task() for managed_node3/reboot 9704 1726776668.40886: done queuing things up, now waiting for results queue to drain 9704 1726776668.40888: waiting for pending results... 10432 1726776668.41007: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems 10432 1726776668.41112: in run() - task 120fa90a-8a95-fee7-0cea-000000000402 10432 1726776668.41129: variable 'ansible_search_path' from source: unknown 10432 1726776668.41134: variable 'ansible_search_path' from source: unknown 10432 1726776668.41162: calling self._execute() 10432 1726776668.41223: variable 'ansible_host' from source: host vars for 'managed_node3' 10432 1726776668.41232: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10432 1726776668.41243: variable 'omit' from source: magic vars 10432 1726776668.41620: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10432 1726776668.43057: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10432 1726776668.43099: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10432 1726776668.43137: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10432 1726776668.43166: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10432 1726776668.43188: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10432 1726776668.43242: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10432 1726776668.43263: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10432 1726776668.43283: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10432 1726776668.43311: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10432 1726776668.43322: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10432 1726776668.43398: variable '__kernel_settings_is_transactional' from source: set_fact 10432 1726776668.43414: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 10432 1726776668.43418: when evaluation is False, skipping this task 10432 1726776668.43421: _execute() done 10432 1726776668.43425: dumping result to json 10432 1726776668.43430: done dumping result, returning 10432 1726776668.43436: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems [120fa90a-8a95-fee7-0cea-000000000402] 10432 1726776668.43444: sending task result for task 120fa90a-8a95-fee7-0cea-000000000402 10432 1726776668.43465: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000402 10432 1726776668.43469: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_is_transactional | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776668.43566: no more pending results, returning what we have 9704 1726776668.43569: results queue empty 9704 1726776668.43569: checking for any_errors_fatal 9704 1726776668.43574: done checking for any_errors_fatal 9704 1726776668.43575: checking for max_fail_percentage 9704 1726776668.43576: done checking for max_fail_percentage 9704 1726776668.43577: checking to see if all hosts have failed and the running result is not ok 9704 1726776668.43577: done checking to see if all hosts have failed 9704 1726776668.43578: getting the remaining hosts for this loop 9704 1726776668.43579: done getting the remaining hosts for this loop 9704 1726776668.43582: getting the next task for host managed_node3 9704 1726776668.43587: done getting next task for host managed_node3 9704 1726776668.43590: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set 9704 1726776668.43592: ^ state is: HOST STATE: block=2, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776668.43606: getting variables 9704 1726776668.43607: in VariableManager get_vars() 9704 1726776668.43637: Calling all_inventory to load vars for managed_node3 9704 1726776668.43641: Calling groups_inventory to load vars for managed_node3 9704 1726776668.43643: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776668.43650: Calling all_plugins_play to load vars for managed_node3 9704 1726776668.43653: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776668.43655: Calling groups_plugins_play to load vars for managed_node3 9704 1726776668.43800: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776668.43913: done with get_vars() 9704 1726776668.43919: done getting variables 9704 1726776668.43957: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:34 Thursday 19 September 2024 16:11:08 -0400 (0:00:00.032) 0:00:21.368 **** 9704 1726776668.43976: entering _queue_task() for managed_node3/fail 9704 1726776668.44121: worker is 1 (out of 1 available) 9704 1726776668.44135: exiting _queue_task() for managed_node3/fail 9704 1726776668.44146: done queuing things up, now waiting for results queue to drain 9704 1726776668.44149: waiting for pending results... 10433 1726776668.44264: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set 10433 1726776668.44371: in run() - task 120fa90a-8a95-fee7-0cea-000000000403 10433 1726776668.44386: variable 'ansible_search_path' from source: unknown 10433 1726776668.44390: variable 'ansible_search_path' from source: unknown 10433 1726776668.44416: calling self._execute() 10433 1726776668.44479: variable 'ansible_host' from source: host vars for 'managed_node3' 10433 1726776668.44487: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10433 1726776668.44495: variable 'omit' from source: magic vars 10433 1726776668.44819: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10433 1726776668.46299: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10433 1726776668.46346: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10433 1726776668.46373: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10433 1726776668.46402: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10433 1726776668.46423: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10433 1726776668.46479: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10433 1726776668.46500: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10433 1726776668.46518: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10433 1726776668.46553: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10433 1726776668.46565: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10433 1726776668.46643: variable '__kernel_settings_is_transactional' from source: set_fact 10433 1726776668.46658: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 10433 1726776668.46663: when evaluation is False, skipping this task 10433 1726776668.46666: _execute() done 10433 1726776668.46670: dumping result to json 10433 1726776668.46673: done dumping result, returning 10433 1726776668.46679: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set [120fa90a-8a95-fee7-0cea-000000000403] 10433 1726776668.46685: sending task result for task 120fa90a-8a95-fee7-0cea-000000000403 10433 1726776668.46707: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000403 10433 1726776668.46711: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_is_transactional | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776668.46811: no more pending results, returning what we have 9704 1726776668.46814: results queue empty 9704 1726776668.46815: checking for any_errors_fatal 9704 1726776668.46820: done checking for any_errors_fatal 9704 1726776668.46821: checking for max_fail_percentage 9704 1726776668.46823: done checking for max_fail_percentage 9704 1726776668.46823: checking to see if all hosts have failed and the running result is not ok 9704 1726776668.46824: done checking to see if all hosts have failed 9704 1726776668.46824: getting the remaining hosts for this loop 9704 1726776668.46825: done getting the remaining hosts for this loop 9704 1726776668.46830: getting the next task for host managed_node3 9704 1726776668.46837: done getting next task for host managed_node3 9704 1726776668.46841: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config 9704 1726776668.46843: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776668.46856: getting variables 9704 1726776668.46857: in VariableManager get_vars() 9704 1726776668.46888: Calling all_inventory to load vars for managed_node3 9704 1726776668.46891: Calling groups_inventory to load vars for managed_node3 9704 1726776668.46893: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776668.46900: Calling all_plugins_play to load vars for managed_node3 9704 1726776668.46902: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776668.46905: Calling groups_plugins_play to load vars for managed_node3 9704 1726776668.47023: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776668.47144: done with get_vars() 9704 1726776668.47152: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Read tuned main config] ****** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:42 Thursday 19 September 2024 16:11:08 -0400 (0:00:00.032) 0:00:21.400 **** 9704 1726776668.47204: entering _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776668.47364: worker is 1 (out of 1 available) 9704 1726776668.47379: exiting _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776668.47391: done queuing things up, now waiting for results queue to drain 9704 1726776668.47393: waiting for pending results... 10434 1726776668.47509: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config 10434 1726776668.47617: in run() - task 120fa90a-8a95-fee7-0cea-000000000405 10434 1726776668.47635: variable 'ansible_search_path' from source: unknown 10434 1726776668.47641: variable 'ansible_search_path' from source: unknown 10434 1726776668.47669: calling self._execute() 10434 1726776668.47730: variable 'ansible_host' from source: host vars for 'managed_node3' 10434 1726776668.47741: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10434 1726776668.47749: variable 'omit' from source: magic vars 10434 1726776668.47819: variable 'omit' from source: magic vars 10434 1726776668.47855: variable 'omit' from source: magic vars 10434 1726776668.47878: variable '__kernel_settings_tuned_main_conf_file' from source: role '' all vars 10434 1726776668.48143: variable '__kernel_settings_tuned_main_conf_file' from source: role '' all vars 10434 1726776668.48202: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10434 1726776668.48230: variable 'omit' from source: magic vars 10434 1726776668.48262: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10434 1726776668.48286: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10434 1726776668.48301: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10434 1726776668.48313: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10434 1726776668.48321: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10434 1726776668.48349: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10434 1726776668.48354: variable 'ansible_host' from source: host vars for 'managed_node3' 10434 1726776668.48359: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10434 1726776668.48422: Set connection var ansible_module_compression to ZIP_DEFLATED 10434 1726776668.48431: Set connection var ansible_shell_type to sh 10434 1726776668.48441: Set connection var ansible_timeout to 10 10434 1726776668.48447: Set connection var ansible_pipelining to False 10434 1726776668.48454: Set connection var ansible_shell_executable to /bin/sh 10434 1726776668.48457: Set connection var ansible_connection to ssh 10434 1726776668.48471: variable 'ansible_shell_executable' from source: unknown 10434 1726776668.48475: variable 'ansible_connection' from source: unknown 10434 1726776668.48478: variable 'ansible_module_compression' from source: unknown 10434 1726776668.48481: variable 'ansible_shell_type' from source: unknown 10434 1726776668.48484: variable 'ansible_shell_executable' from source: unknown 10434 1726776668.48486: variable 'ansible_host' from source: host vars for 'managed_node3' 10434 1726776668.48488: variable 'ansible_pipelining' from source: unknown 10434 1726776668.48490: variable 'ansible_timeout' from source: unknown 10434 1726776668.48492: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10434 1726776668.48610: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10434 1726776668.48620: variable 'omit' from source: magic vars 10434 1726776668.48625: starting attempt loop 10434 1726776668.48630: running the handler 10434 1726776668.48644: _low_level_execute_command(): starting 10434 1726776668.48652: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10434 1726776668.50971: stdout chunk (state=2): >>>/root <<< 10434 1726776668.51090: stderr chunk (state=3): >>><<< 10434 1726776668.51096: stdout chunk (state=3): >>><<< 10434 1726776668.51112: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10434 1726776668.51124: _low_level_execute_command(): starting 10434 1726776668.51131: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502 `" && echo ansible-tmp-1726776668.5111892-10434-103566107000502="` echo /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502 `" ) && sleep 0' 10434 1726776668.53585: stdout chunk (state=2): >>>ansible-tmp-1726776668.5111892-10434-103566107000502=/root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502 <<< 10434 1726776668.53709: stderr chunk (state=3): >>><<< 10434 1726776668.53715: stdout chunk (state=3): >>><<< 10434 1726776668.53730: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776668.5111892-10434-103566107000502=/root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502 , stderr= 10434 1726776668.53770: variable 'ansible_module_compression' from source: unknown 10434 1726776668.53800: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.kernel_settings_get_config-ZIP_DEFLATED 10434 1726776668.53831: variable 'ansible_facts' from source: unknown 10434 1726776668.53897: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/AnsiballZ_kernel_settings_get_config.py 10434 1726776668.53994: Sending initial data 10434 1726776668.54001: Sent initial data (174 bytes) 10434 1726776668.56495: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpda7_sfi7 /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/AnsiballZ_kernel_settings_get_config.py <<< 10434 1726776668.57458: stderr chunk (state=3): >>><<< 10434 1726776668.57465: stdout chunk (state=3): >>><<< 10434 1726776668.57481: done transferring module to remote 10434 1726776668.57490: _low_level_execute_command(): starting 10434 1726776668.57495: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/ /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10434 1726776668.59835: stderr chunk (state=2): >>><<< 10434 1726776668.59844: stdout chunk (state=2): >>><<< 10434 1726776668.59856: _low_level_execute_command() done: rc=0, stdout=, stderr= 10434 1726776668.59860: _low_level_execute_command(): starting 10434 1726776668.59865: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10434 1726776668.75708: stdout chunk (state=2): >>> {"changed": false, "data": {"daemon": "1", "dynamic_tuning": "0", "sleep_interval": "1", "update_interval": "10", "recommend_command": "1", "reapply_sysctl": "1", "default_instance_priority": "0", "udev_buffer_size": "1MB", "log_file_count": "2", "log_file_max_size": "1MB"}, "invocation": {"module_args": {"path": "/etc/tuned/tuned-main.conf"}}} <<< 10434 1726776668.76802: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10434 1726776668.76852: stderr chunk (state=3): >>><<< 10434 1726776668.76859: stdout chunk (state=3): >>><<< 10434 1726776668.76874: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "data": {"daemon": "1", "dynamic_tuning": "0", "sleep_interval": "1", "update_interval": "10", "recommend_command": "1", "reapply_sysctl": "1", "default_instance_priority": "0", "udev_buffer_size": "1MB", "log_file_count": "2", "log_file_max_size": "1MB"}, "invocation": {"module_args": {"path": "/etc/tuned/tuned-main.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 10434 1726776668.76901: done with _execute_module (fedora.linux_system_roles.kernel_settings_get_config, {'path': '/etc/tuned/tuned-main.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'fedora.linux_system_roles.kernel_settings_get_config', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10434 1726776668.76912: _low_level_execute_command(): starting 10434 1726776668.76918: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776668.5111892-10434-103566107000502/ > /dev/null 2>&1 && sleep 0' 10434 1726776668.79319: stderr chunk (state=2): >>><<< 10434 1726776668.79327: stdout chunk (state=2): >>><<< 10434 1726776668.79341: _low_level_execute_command() done: rc=0, stdout=, stderr= 10434 1726776668.79348: handler run complete 10434 1726776668.79362: attempt loop complete, returning result 10434 1726776668.79366: _execute() done 10434 1726776668.79370: dumping result to json 10434 1726776668.79374: done dumping result, returning 10434 1726776668.79381: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config [120fa90a-8a95-fee7-0cea-000000000405] 10434 1726776668.79387: sending task result for task 120fa90a-8a95-fee7-0cea-000000000405 10434 1726776668.79418: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000405 10434 1726776668.79423: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "data": { "daemon": "1", "default_instance_priority": "0", "dynamic_tuning": "0", "log_file_count": "2", "log_file_max_size": "1MB", "reapply_sysctl": "1", "recommend_command": "1", "sleep_interval": "1", "udev_buffer_size": "1MB", "update_interval": "10" } } 9704 1726776668.79588: no more pending results, returning what we have 9704 1726776668.79591: results queue empty 9704 1726776668.79592: checking for any_errors_fatal 9704 1726776668.79598: done checking for any_errors_fatal 9704 1726776668.79598: checking for max_fail_percentage 9704 1726776668.79599: done checking for max_fail_percentage 9704 1726776668.79600: checking to see if all hosts have failed and the running result is not ok 9704 1726776668.79600: done checking to see if all hosts have failed 9704 1726776668.79601: getting the remaining hosts for this loop 9704 1726776668.79602: done getting the remaining hosts for this loop 9704 1726776668.79605: getting the next task for host managed_node3 9704 1726776668.79610: done getting next task for host managed_node3 9704 1726776668.79613: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory 9704 1726776668.79615: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776668.79625: getting variables 9704 1726776668.79626: in VariableManager get_vars() 9704 1726776668.79662: Calling all_inventory to load vars for managed_node3 9704 1726776668.79665: Calling groups_inventory to load vars for managed_node3 9704 1726776668.79667: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776668.79674: Calling all_plugins_play to load vars for managed_node3 9704 1726776668.79676: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776668.79677: Calling groups_plugins_play to load vars for managed_node3 9704 1726776668.79825: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776668.79948: done with get_vars() 9704 1726776668.79955: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:50 Thursday 19 September 2024 16:11:08 -0400 (0:00:00.328) 0:00:21.728 **** 9704 1726776668.80017: entering _queue_task() for managed_node3/stat 9704 1726776668.80187: worker is 1 (out of 1 available) 9704 1726776668.80202: exiting _queue_task() for managed_node3/stat 9704 1726776668.80214: done queuing things up, now waiting for results queue to drain 9704 1726776668.80217: waiting for pending results... 10445 1726776668.80335: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory 10445 1726776668.80446: in run() - task 120fa90a-8a95-fee7-0cea-000000000406 10445 1726776668.80462: variable 'ansible_search_path' from source: unknown 10445 1726776668.80465: variable 'ansible_search_path' from source: unknown 10445 1726776668.80500: variable '__prof_from_conf' from source: task vars 10445 1726776668.80727: variable '__prof_from_conf' from source: task vars 10445 1726776668.80861: variable '__data' from source: task vars 10445 1726776668.80915: variable '__kernel_settings_register_tuned_main' from source: set_fact 10445 1726776668.81054: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10445 1726776668.81064: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10445 1726776668.81108: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10445 1726776668.81124: variable 'omit' from source: magic vars 10445 1726776668.81205: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776668.81214: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776668.81220: variable 'omit' from source: magic vars 10445 1726776668.81394: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10445 1726776668.82923: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10445 1726776668.82975: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10445 1726776668.83006: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10445 1726776668.83035: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10445 1726776668.83066: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10445 1726776668.83122: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10445 1726776668.83148: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10445 1726776668.83168: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10445 1726776668.83195: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10445 1726776668.83207: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10445 1726776668.83294: variable 'item' from source: unknown 10445 1726776668.83307: Evaluated conditional (item | length > 0): False 10445 1726776668.83312: when evaluation is False, skipping this task 10445 1726776668.83337: variable 'item' from source: unknown 10445 1726776668.83389: variable 'item' from source: unknown skipping: [managed_node3] => (item=) => { "ansible_loop_var": "item", "changed": false, "false_condition": "item | length > 0", "item": "", "skip_reason": "Conditional result was False" } 10445 1726776668.83475: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776668.83485: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776668.83494: variable 'omit' from source: magic vars 10445 1726776668.83618: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10445 1726776668.83639: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10445 1726776668.83657: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10445 1726776668.83683: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10445 1726776668.83695: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10445 1726776668.83750: variable 'item' from source: unknown 10445 1726776668.83759: Evaluated conditional (item | length > 0): True 10445 1726776668.83765: variable 'omit' from source: magic vars 10445 1726776668.83792: variable 'omit' from source: magic vars 10445 1726776668.83824: variable 'item' from source: unknown 10445 1726776668.83869: variable 'item' from source: unknown 10445 1726776668.83883: variable 'omit' from source: magic vars 10445 1726776668.83902: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10445 1726776668.83924: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10445 1726776668.83941: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10445 1726776668.83955: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10445 1726776668.83964: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10445 1726776668.83986: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10445 1726776668.83990: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776668.83994: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776668.84060: Set connection var ansible_module_compression to ZIP_DEFLATED 10445 1726776668.84068: Set connection var ansible_shell_type to sh 10445 1726776668.84076: Set connection var ansible_timeout to 10 10445 1726776668.84081: Set connection var ansible_pipelining to False 10445 1726776668.84088: Set connection var ansible_shell_executable to /bin/sh 10445 1726776668.84091: Set connection var ansible_connection to ssh 10445 1726776668.84104: variable 'ansible_shell_executable' from source: unknown 10445 1726776668.84107: variable 'ansible_connection' from source: unknown 10445 1726776668.84110: variable 'ansible_module_compression' from source: unknown 10445 1726776668.84113: variable 'ansible_shell_type' from source: unknown 10445 1726776668.84117: variable 'ansible_shell_executable' from source: unknown 10445 1726776668.84120: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776668.84124: variable 'ansible_pipelining' from source: unknown 10445 1726776668.84127: variable 'ansible_timeout' from source: unknown 10445 1726776668.84133: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776668.84218: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10445 1726776668.84231: variable 'omit' from source: magic vars 10445 1726776668.84237: starting attempt loop 10445 1726776668.84241: running the handler 10445 1726776668.84252: _low_level_execute_command(): starting 10445 1726776668.84261: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10445 1726776668.86561: stdout chunk (state=2): >>>/root <<< 10445 1726776668.86695: stderr chunk (state=3): >>><<< 10445 1726776668.86701: stdout chunk (state=3): >>><<< 10445 1726776668.86719: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10445 1726776668.86732: _low_level_execute_command(): starting 10445 1726776668.86737: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272 `" && echo ansible-tmp-1726776668.8672657-10445-208898276663272="` echo /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272 `" ) && sleep 0' 10445 1726776668.89187: stdout chunk (state=2): >>>ansible-tmp-1726776668.8672657-10445-208898276663272=/root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272 <<< 10445 1726776668.89318: stderr chunk (state=3): >>><<< 10445 1726776668.89324: stdout chunk (state=3): >>><<< 10445 1726776668.89342: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776668.8672657-10445-208898276663272=/root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272 , stderr= 10445 1726776668.89374: variable 'ansible_module_compression' from source: unknown 10445 1726776668.89416: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10445 1726776668.89447: variable 'ansible_facts' from source: unknown 10445 1726776668.89512: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/AnsiballZ_stat.py 10445 1726776668.89606: Sending initial data 10445 1726776668.89612: Sent initial data (152 bytes) 10445 1726776668.92115: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp0a0v4kah /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/AnsiballZ_stat.py <<< 10445 1726776668.93080: stderr chunk (state=3): >>><<< 10445 1726776668.93086: stdout chunk (state=3): >>><<< 10445 1726776668.93102: done transferring module to remote 10445 1726776668.93111: _low_level_execute_command(): starting 10445 1726776668.93116: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/ /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/AnsiballZ_stat.py && sleep 0' 10445 1726776668.95473: stderr chunk (state=2): >>><<< 10445 1726776668.95479: stdout chunk (state=2): >>><<< 10445 1726776668.95491: _low_level_execute_command() done: rc=0, stdout=, stderr= 10445 1726776668.95495: _low_level_execute_command(): starting 10445 1726776668.95500: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/AnsiballZ_stat.py && sleep 0' 10445 1726776669.10754: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/profiles", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 10445 1726776669.11865: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10445 1726776669.11912: stderr chunk (state=3): >>><<< 10445 1726776669.11918: stdout chunk (state=3): >>><<< 10445 1726776669.11936: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/profiles", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 10445 1726776669.11958: done with _execute_module (stat, {'path': '/etc/tuned/profiles', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10445 1726776669.11970: _low_level_execute_command(): starting 10445 1726776669.11976: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776668.8672657-10445-208898276663272/ > /dev/null 2>&1 && sleep 0' 10445 1726776669.14422: stderr chunk (state=2): >>><<< 10445 1726776669.14430: stdout chunk (state=2): >>><<< 10445 1726776669.14444: _low_level_execute_command() done: rc=0, stdout=, stderr= 10445 1726776669.14451: handler run complete 10445 1726776669.14466: attempt loop complete, returning result 10445 1726776669.14481: variable 'item' from source: unknown 10445 1726776669.14546: variable 'item' from source: unknown ok: [managed_node3] => (item=/etc/tuned/profiles) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/tuned/profiles", "stat": { "exists": false } } 10445 1726776669.14634: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776669.14644: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776669.14654: variable 'omit' from source: magic vars 10445 1726776669.14761: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10445 1726776669.14784: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10445 1726776669.14803: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10445 1726776669.14831: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10445 1726776669.14844: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10445 1726776669.14901: variable 'item' from source: unknown 10445 1726776669.14911: Evaluated conditional (item | length > 0): True 10445 1726776669.14916: variable 'omit' from source: magic vars 10445 1726776669.14928: variable 'omit' from source: magic vars 10445 1726776669.14960: variable 'item' from source: unknown 10445 1726776669.15006: variable 'item' from source: unknown 10445 1726776669.15019: variable 'omit' from source: magic vars 10445 1726776669.15037: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10445 1726776669.15049: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10445 1726776669.15055: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10445 1726776669.15067: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10445 1726776669.15071: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776669.15075: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776669.15124: Set connection var ansible_module_compression to ZIP_DEFLATED 10445 1726776669.15132: Set connection var ansible_shell_type to sh 10445 1726776669.15141: Set connection var ansible_timeout to 10 10445 1726776669.15146: Set connection var ansible_pipelining to False 10445 1726776669.15153: Set connection var ansible_shell_executable to /bin/sh 10445 1726776669.15156: Set connection var ansible_connection to ssh 10445 1726776669.15169: variable 'ansible_shell_executable' from source: unknown 10445 1726776669.15172: variable 'ansible_connection' from source: unknown 10445 1726776669.15175: variable 'ansible_module_compression' from source: unknown 10445 1726776669.15178: variable 'ansible_shell_type' from source: unknown 10445 1726776669.15182: variable 'ansible_shell_executable' from source: unknown 10445 1726776669.15185: variable 'ansible_host' from source: host vars for 'managed_node3' 10445 1726776669.15189: variable 'ansible_pipelining' from source: unknown 10445 1726776669.15192: variable 'ansible_timeout' from source: unknown 10445 1726776669.15196: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10445 1726776669.15268: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10445 1726776669.15278: variable 'omit' from source: magic vars 10445 1726776669.15283: starting attempt loop 10445 1726776669.15287: running the handler 10445 1726776669.15294: _low_level_execute_command(): starting 10445 1726776669.15298: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10445 1726776669.17493: stdout chunk (state=2): >>>/root <<< 10445 1726776669.17610: stderr chunk (state=3): >>><<< 10445 1726776669.17616: stdout chunk (state=3): >>><<< 10445 1726776669.17631: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10445 1726776669.17642: _low_level_execute_command(): starting 10445 1726776669.17647: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802 `" && echo ansible-tmp-1726776669.1763682-10445-148052934174802="` echo /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802 `" ) && sleep 0' 10445 1726776669.20099: stdout chunk (state=2): >>>ansible-tmp-1726776669.1763682-10445-148052934174802=/root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802 <<< 10445 1726776669.20223: stderr chunk (state=3): >>><<< 10445 1726776669.20231: stdout chunk (state=3): >>><<< 10445 1726776669.20245: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776669.1763682-10445-148052934174802=/root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802 , stderr= 10445 1726776669.20275: variable 'ansible_module_compression' from source: unknown 10445 1726776669.20310: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10445 1726776669.20325: variable 'ansible_facts' from source: unknown 10445 1726776669.20384: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/AnsiballZ_stat.py 10445 1726776669.20468: Sending initial data 10445 1726776669.20476: Sent initial data (152 bytes) 10445 1726776669.22956: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpzblzi2ff /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/AnsiballZ_stat.py <<< 10445 1726776669.23897: stderr chunk (state=3): >>><<< 10445 1726776669.23903: stdout chunk (state=3): >>><<< 10445 1726776669.23918: done transferring module to remote 10445 1726776669.23926: _low_level_execute_command(): starting 10445 1726776669.23933: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/ /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/AnsiballZ_stat.py && sleep 0' 10445 1726776669.26287: stderr chunk (state=2): >>><<< 10445 1726776669.26296: stdout chunk (state=2): >>><<< 10445 1726776669.26311: _low_level_execute_command() done: rc=0, stdout=, stderr= 10445 1726776669.26315: _low_level_execute_command(): starting 10445 1726776669.26320: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/AnsiballZ_stat.py && sleep 0' 10445 1726776669.42249: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned", "mode": "0755", "isdir": true, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 159, "inode": 917919, "dev": 51713, "nlink": 4, "atime": 1726776658.090236, "mtime": 1726776656.4062142, "ctime": 1726776656.4062142, "wusr": true, "rusr": true, "xusr": true, "wgrp": false, "rgrp": true, "xgrp": true, "woth": false, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "pw_name": "root", "gr_name": "root", "mimetype": "inode/directory", "charset": "binary", "version": "1785990601", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 10445 1726776669.43455: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10445 1726776669.43503: stderr chunk (state=3): >>><<< 10445 1726776669.43510: stdout chunk (state=3): >>><<< 10445 1726776669.43526: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned", "mode": "0755", "isdir": true, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 159, "inode": 917919, "dev": 51713, "nlink": 4, "atime": 1726776658.090236, "mtime": 1726776656.4062142, "ctime": 1726776656.4062142, "wusr": true, "rusr": true, "xusr": true, "wgrp": false, "rgrp": true, "xgrp": true, "woth": false, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "pw_name": "root", "gr_name": "root", "mimetype": "inode/directory", "charset": "binary", "version": "1785990601", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 10445 1726776669.43562: done with _execute_module (stat, {'path': '/etc/tuned', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10445 1726776669.43571: _low_level_execute_command(): starting 10445 1726776669.43577: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776669.1763682-10445-148052934174802/ > /dev/null 2>&1 && sleep 0' 10445 1726776669.46020: stderr chunk (state=2): >>><<< 10445 1726776669.46027: stdout chunk (state=2): >>><<< 10445 1726776669.46042: _low_level_execute_command() done: rc=0, stdout=, stderr= 10445 1726776669.46048: handler run complete 10445 1726776669.46079: attempt loop complete, returning result 10445 1726776669.46094: variable 'item' from source: unknown 10445 1726776669.46156: variable 'item' from source: unknown ok: [managed_node3] => (item=/etc/tuned) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/tuned", "stat": { "atime": 1726776658.090236, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1726776656.4062142, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 917919, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/directory", "mode": "0755", "mtime": 1726776656.4062142, "nlink": 4, "path": "/etc/tuned", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 159, "uid": 0, "version": "1785990601", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 10445 1726776669.46199: dumping result to json 10445 1726776669.46209: done dumping result, returning 10445 1726776669.46216: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory [120fa90a-8a95-fee7-0cea-000000000406] 10445 1726776669.46222: sending task result for task 120fa90a-8a95-fee7-0cea-000000000406 10445 1726776669.46264: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000406 10445 1726776669.46268: WORKER PROCESS EXITING 9704 1726776669.46514: no more pending results, returning what we have 9704 1726776669.46517: results queue empty 9704 1726776669.46518: checking for any_errors_fatal 9704 1726776669.46522: done checking for any_errors_fatal 9704 1726776669.46523: checking for max_fail_percentage 9704 1726776669.46524: done checking for max_fail_percentage 9704 1726776669.46525: checking to see if all hosts have failed and the running result is not ok 9704 1726776669.46525: done checking to see if all hosts have failed 9704 1726776669.46526: getting the remaining hosts for this loop 9704 1726776669.46527: done getting the remaining hosts for this loop 9704 1726776669.46531: getting the next task for host managed_node3 9704 1726776669.46535: done getting next task for host managed_node3 9704 1726776669.46538: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir 9704 1726776669.46542: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776669.46551: getting variables 9704 1726776669.46552: in VariableManager get_vars() 9704 1726776669.46574: Calling all_inventory to load vars for managed_node3 9704 1726776669.46576: Calling groups_inventory to load vars for managed_node3 9704 1726776669.46577: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776669.46582: Calling all_plugins_play to load vars for managed_node3 9704 1726776669.46584: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776669.46585: Calling groups_plugins_play to load vars for managed_node3 9704 1726776669.46686: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776669.46800: done with get_vars() 9704 1726776669.46808: done getting variables 9704 1726776669.46850: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:63 Thursday 19 September 2024 16:11:09 -0400 (0:00:00.668) 0:00:22.397 **** 9704 1726776669.46871: entering _queue_task() for managed_node3/set_fact 9704 1726776669.47035: worker is 1 (out of 1 available) 9704 1726776669.47053: exiting _queue_task() for managed_node3/set_fact 9704 1726776669.47065: done queuing things up, now waiting for results queue to drain 9704 1726776669.47067: waiting for pending results... 10460 1726776669.47183: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir 10460 1726776669.47298: in run() - task 120fa90a-8a95-fee7-0cea-000000000407 10460 1726776669.47315: variable 'ansible_search_path' from source: unknown 10460 1726776669.47319: variable 'ansible_search_path' from source: unknown 10460 1726776669.47347: calling self._execute() 10460 1726776669.47413: variable 'ansible_host' from source: host vars for 'managed_node3' 10460 1726776669.47421: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10460 1726776669.47431: variable 'omit' from source: magic vars 10460 1726776669.47573: variable 'omit' from source: magic vars 10460 1726776669.47610: variable 'omit' from source: magic vars 10460 1726776669.47911: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10460 1726776669.49363: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10460 1726776669.49414: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10460 1726776669.49445: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10460 1726776669.49473: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10460 1726776669.49494: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10460 1726776669.49551: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10460 1726776669.49575: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10460 1726776669.49594: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10460 1726776669.49620: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10460 1726776669.49633: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10460 1726776669.49665: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10460 1726776669.49685: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10460 1726776669.49703: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10460 1726776669.49730: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10460 1726776669.49742: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10460 1726776669.49780: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10460 1726776669.49798: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10460 1726776669.49812: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10460 1726776669.49836: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10460 1726776669.49845: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10460 1726776669.49988: variable '__kernel_settings_find_profile_dirs' from source: set_fact 10460 1726776669.50055: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10460 1726776669.50161: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10460 1726776669.50185: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10460 1726776669.50205: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10460 1726776669.50224: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10460 1726776669.50254: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10460 1726776669.50268: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10460 1726776669.50282: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10460 1726776669.50296: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10460 1726776669.50331: variable 'omit' from source: magic vars 10460 1726776669.50357: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10460 1726776669.50376: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10460 1726776669.50391: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10460 1726776669.50404: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10460 1726776669.50414: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10460 1726776669.50438: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10460 1726776669.50443: variable 'ansible_host' from source: host vars for 'managed_node3' 10460 1726776669.50447: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10460 1726776669.50505: Set connection var ansible_module_compression to ZIP_DEFLATED 10460 1726776669.50510: Set connection var ansible_shell_type to sh 10460 1726776669.50515: Set connection var ansible_timeout to 10 10460 1726776669.50518: Set connection var ansible_pipelining to False 10460 1726776669.50523: Set connection var ansible_shell_executable to /bin/sh 10460 1726776669.50524: Set connection var ansible_connection to ssh 10460 1726776669.50540: variable 'ansible_shell_executable' from source: unknown 10460 1726776669.50543: variable 'ansible_connection' from source: unknown 10460 1726776669.50545: variable 'ansible_module_compression' from source: unknown 10460 1726776669.50547: variable 'ansible_shell_type' from source: unknown 10460 1726776669.50549: variable 'ansible_shell_executable' from source: unknown 10460 1726776669.50551: variable 'ansible_host' from source: host vars for 'managed_node3' 10460 1726776669.50555: variable 'ansible_pipelining' from source: unknown 10460 1726776669.50557: variable 'ansible_timeout' from source: unknown 10460 1726776669.50559: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10460 1726776669.50614: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10460 1726776669.50622: variable 'omit' from source: magic vars 10460 1726776669.50626: starting attempt loop 10460 1726776669.50631: running the handler 10460 1726776669.50639: handler run complete 10460 1726776669.50646: attempt loop complete, returning result 10460 1726776669.50648: _execute() done 10460 1726776669.50650: dumping result to json 10460 1726776669.50652: done dumping result, returning 10460 1726776669.50655: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir [120fa90a-8a95-fee7-0cea-000000000407] 10460 1726776669.50659: sending task result for task 120fa90a-8a95-fee7-0cea-000000000407 10460 1726776669.50677: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000407 10460 1726776669.50679: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_profile_parent": "/etc/tuned" }, "changed": false } 9704 1726776669.50967: no more pending results, returning what we have 9704 1726776669.50970: results queue empty 9704 1726776669.50970: checking for any_errors_fatal 9704 1726776669.50978: done checking for any_errors_fatal 9704 1726776669.50978: checking for max_fail_percentage 9704 1726776669.50979: done checking for max_fail_percentage 9704 1726776669.50980: checking to see if all hosts have failed and the running result is not ok 9704 1726776669.50980: done checking to see if all hosts have failed 9704 1726776669.50981: getting the remaining hosts for this loop 9704 1726776669.50982: done getting the remaining hosts for this loop 9704 1726776669.50986: getting the next task for host managed_node3 9704 1726776669.50991: done getting next task for host managed_node3 9704 1726776669.50993: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started 9704 1726776669.50995: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776669.51004: getting variables 9704 1726776669.51005: in VariableManager get_vars() 9704 1726776669.51031: Calling all_inventory to load vars for managed_node3 9704 1726776669.51033: Calling groups_inventory to load vars for managed_node3 9704 1726776669.51034: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776669.51041: Calling all_plugins_play to load vars for managed_node3 9704 1726776669.51043: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776669.51044: Calling groups_plugins_play to load vars for managed_node3 9704 1726776669.51194: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776669.51309: done with get_vars() 9704 1726776669.51318: done getting variables 9704 1726776669.51359: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:67 Thursday 19 September 2024 16:11:09 -0400 (0:00:00.045) 0:00:22.442 **** 9704 1726776669.51380: entering _queue_task() for managed_node3/service 9704 1726776669.51541: worker is 1 (out of 1 available) 9704 1726776669.51555: exiting _queue_task() for managed_node3/service 9704 1726776669.51566: done queuing things up, now waiting for results queue to drain 9704 1726776669.51568: waiting for pending results... 10461 1726776669.51688: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started 10461 1726776669.51803: in run() - task 120fa90a-8a95-fee7-0cea-000000000408 10461 1726776669.51817: variable 'ansible_search_path' from source: unknown 10461 1726776669.51821: variable 'ansible_search_path' from source: unknown 10461 1726776669.51858: variable '__kernel_settings_services' from source: include_vars 10461 1726776669.52083: variable '__kernel_settings_services' from source: include_vars 10461 1726776669.52143: variable 'omit' from source: magic vars 10461 1726776669.52226: variable 'ansible_host' from source: host vars for 'managed_node3' 10461 1726776669.52238: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10461 1726776669.52249: variable 'omit' from source: magic vars 10461 1726776669.52299: variable 'omit' from source: magic vars 10461 1726776669.52331: variable 'omit' from source: magic vars 10461 1726776669.52366: variable 'item' from source: unknown 10461 1726776669.52417: variable 'item' from source: unknown 10461 1726776669.52438: variable 'omit' from source: magic vars 10461 1726776669.52470: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10461 1726776669.52495: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10461 1726776669.52513: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10461 1726776669.52526: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10461 1726776669.52537: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10461 1726776669.52563: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10461 1726776669.52568: variable 'ansible_host' from source: host vars for 'managed_node3' 10461 1726776669.52573: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10461 1726776669.52637: Set connection var ansible_module_compression to ZIP_DEFLATED 10461 1726776669.52647: Set connection var ansible_shell_type to sh 10461 1726776669.52655: Set connection var ansible_timeout to 10 10461 1726776669.52661: Set connection var ansible_pipelining to False 10461 1726776669.52666: Set connection var ansible_shell_executable to /bin/sh 10461 1726776669.52668: Set connection var ansible_connection to ssh 10461 1726776669.52678: variable 'ansible_shell_executable' from source: unknown 10461 1726776669.52681: variable 'ansible_connection' from source: unknown 10461 1726776669.52683: variable 'ansible_module_compression' from source: unknown 10461 1726776669.52684: variable 'ansible_shell_type' from source: unknown 10461 1726776669.52686: variable 'ansible_shell_executable' from source: unknown 10461 1726776669.52688: variable 'ansible_host' from source: host vars for 'managed_node3' 10461 1726776669.52690: variable 'ansible_pipelining' from source: unknown 10461 1726776669.52691: variable 'ansible_timeout' from source: unknown 10461 1726776669.52693: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10461 1726776669.52794: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10461 1726776669.52806: variable 'omit' from source: magic vars 10461 1726776669.52812: starting attempt loop 10461 1726776669.52815: running the handler 10461 1726776669.52880: variable 'ansible_facts' from source: unknown 10461 1726776669.52963: _low_level_execute_command(): starting 10461 1726776669.52972: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10461 1726776669.55313: stdout chunk (state=2): >>>/root <<< 10461 1726776669.55435: stderr chunk (state=3): >>><<< 10461 1726776669.55444: stdout chunk (state=3): >>><<< 10461 1726776669.55464: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10461 1726776669.55478: _low_level_execute_command(): starting 10461 1726776669.55485: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214 `" && echo ansible-tmp-1726776669.5547252-10461-72244031051214="` echo /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214 `" ) && sleep 0' 10461 1726776669.57960: stdout chunk (state=2): >>>ansible-tmp-1726776669.5547252-10461-72244031051214=/root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214 <<< 10461 1726776669.58086: stderr chunk (state=3): >>><<< 10461 1726776669.58092: stdout chunk (state=3): >>><<< 10461 1726776669.58108: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776669.5547252-10461-72244031051214=/root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214 , stderr= 10461 1726776669.58134: variable 'ansible_module_compression' from source: unknown 10461 1726776669.58174: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 10461 1726776669.58217: variable 'ansible_facts' from source: unknown 10461 1726776669.58372: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/AnsiballZ_systemd.py 10461 1726776669.58499: Sending initial data 10461 1726776669.58506: Sent initial data (154 bytes) 10461 1726776669.60993: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp_aoo3m2b /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/AnsiballZ_systemd.py <<< 10461 1726776669.62809: stderr chunk (state=3): >>><<< 10461 1726776669.62816: stdout chunk (state=3): >>><<< 10461 1726776669.62836: done transferring module to remote 10461 1726776669.62847: _low_level_execute_command(): starting 10461 1726776669.62853: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/ /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/AnsiballZ_systemd.py && sleep 0' 10461 1726776669.65218: stderr chunk (state=2): >>><<< 10461 1726776669.65231: stdout chunk (state=3): >>><<< 10461 1726776669.65243: _low_level_execute_command() done: rc=0, stdout=, stderr= 10461 1726776669.65247: _low_level_execute_command(): starting 10461 1726776669.65252: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/AnsiballZ_systemd.py && sleep 0' 10461 1726776669.93188: stdout chunk (state=2): >>> {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "15002", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15069184", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "Memo<<< 10461 1726776669.93235: stdout chunk (state=3): >>>ryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 10461 1726776669.94916: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10461 1726776669.94967: stderr chunk (state=3): >>><<< 10461 1726776669.94974: stdout chunk (state=3): >>><<< 10461 1726776669.94994: _low_level_execute_command() done: rc=0, stdout= {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "15002", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15069184", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10461 1726776669.95105: done with _execute_module (ansible.legacy.systemd, {'name': 'tuned', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10461 1726776669.95124: _low_level_execute_command(): starting 10461 1726776669.95132: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776669.5547252-10461-72244031051214/ > /dev/null 2>&1 && sleep 0' 10461 1726776669.97555: stderr chunk (state=2): >>><<< 10461 1726776669.97561: stdout chunk (state=2): >>><<< 10461 1726776669.97574: _low_level_execute_command() done: rc=0, stdout=, stderr= 10461 1726776669.97581: handler run complete 10461 1726776669.97612: attempt loop complete, returning result 10461 1726776669.97630: variable 'item' from source: unknown 10461 1726776669.97691: variable 'item' from source: unknown ok: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": false, "enabled": true, "item": "tuned", "name": "tuned", "state": "started", "status": { "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "ActiveState": "active", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "com.redhat.tuned", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "ConfigurationDirectoryMode": "0755", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "ControlGroup": "/system.slice/tuned.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Dynamic System Tuning Daemon", "DevicePolicy": "auto", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "tuned.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "15002", "MemoryAccounting": "yes", "MemoryCurrent": "15069184", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "tuned.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/tuned/tuned.pid", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "WatchdogUSec": "0" } } 10461 1726776669.97790: dumping result to json 10461 1726776669.97809: done dumping result, returning 10461 1726776669.97817: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started [120fa90a-8a95-fee7-0cea-000000000408] 10461 1726776669.97824: sending task result for task 120fa90a-8a95-fee7-0cea-000000000408 10461 1726776669.97934: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000408 10461 1726776669.97939: WORKER PROCESS EXITING 9704 1726776669.98280: no more pending results, returning what we have 9704 1726776669.98283: results queue empty 9704 1726776669.98283: checking for any_errors_fatal 9704 1726776669.98286: done checking for any_errors_fatal 9704 1726776669.98287: checking for max_fail_percentage 9704 1726776669.98288: done checking for max_fail_percentage 9704 1726776669.98288: checking to see if all hosts have failed and the running result is not ok 9704 1726776669.98289: done checking to see if all hosts have failed 9704 1726776669.98289: getting the remaining hosts for this loop 9704 1726776669.98290: done getting the remaining hosts for this loop 9704 1726776669.98292: getting the next task for host managed_node3 9704 1726776669.98296: done getting next task for host managed_node3 9704 1726776669.98299: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists 9704 1726776669.98300: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776669.98307: getting variables 9704 1726776669.98308: in VariableManager get_vars() 9704 1726776669.98331: Calling all_inventory to load vars for managed_node3 9704 1726776669.98333: Calling groups_inventory to load vars for managed_node3 9704 1726776669.98334: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776669.98343: Calling all_plugins_play to load vars for managed_node3 9704 1726776669.98345: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776669.98347: Calling groups_plugins_play to load vars for managed_node3 9704 1726776669.98462: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776669.98580: done with get_vars() 9704 1726776669.98588: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:74 Thursday 19 September 2024 16:11:09 -0400 (0:00:00.472) 0:00:22.915 **** 9704 1726776669.98655: entering _queue_task() for managed_node3/file 9704 1726776669.98813: worker is 1 (out of 1 available) 9704 1726776669.98830: exiting _queue_task() for managed_node3/file 9704 1726776669.98844: done queuing things up, now waiting for results queue to drain 9704 1726776669.98846: waiting for pending results... 10472 1726776669.98963: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists 10472 1726776669.99069: in run() - task 120fa90a-8a95-fee7-0cea-000000000409 10472 1726776669.99083: variable 'ansible_search_path' from source: unknown 10472 1726776669.99087: variable 'ansible_search_path' from source: unknown 10472 1726776669.99114: calling self._execute() 10472 1726776669.99182: variable 'ansible_host' from source: host vars for 'managed_node3' 10472 1726776669.99190: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10472 1726776669.99199: variable 'omit' from source: magic vars 10472 1726776669.99272: variable 'omit' from source: magic vars 10472 1726776669.99307: variable 'omit' from source: magic vars 10472 1726776669.99328: variable '__kernel_settings_profile_dir' from source: role '' all vars 10472 1726776669.99535: variable '__kernel_settings_profile_dir' from source: role '' all vars 10472 1726776669.99604: variable '__kernel_settings_profile_parent' from source: set_fact 10472 1726776669.99611: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10472 1726776669.99640: variable 'omit' from source: magic vars 10472 1726776669.99669: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10472 1726776669.99692: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10472 1726776669.99708: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10472 1726776669.99720: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10472 1726776669.99730: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10472 1726776669.99752: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10472 1726776669.99756: variable 'ansible_host' from source: host vars for 'managed_node3' 10472 1726776669.99758: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10472 1726776669.99819: Set connection var ansible_module_compression to ZIP_DEFLATED 10472 1726776669.99826: Set connection var ansible_shell_type to sh 10472 1726776669.99854: Set connection var ansible_timeout to 10 10472 1726776669.99860: Set connection var ansible_pipelining to False 10472 1726776669.99867: Set connection var ansible_shell_executable to /bin/sh 10472 1726776669.99870: Set connection var ansible_connection to ssh 10472 1726776669.99884: variable 'ansible_shell_executable' from source: unknown 10472 1726776669.99888: variable 'ansible_connection' from source: unknown 10472 1726776669.99892: variable 'ansible_module_compression' from source: unknown 10472 1726776669.99896: variable 'ansible_shell_type' from source: unknown 10472 1726776669.99899: variable 'ansible_shell_executable' from source: unknown 10472 1726776669.99903: variable 'ansible_host' from source: host vars for 'managed_node3' 10472 1726776669.99907: variable 'ansible_pipelining' from source: unknown 10472 1726776669.99910: variable 'ansible_timeout' from source: unknown 10472 1726776669.99914: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10472 1726776670.00046: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10472 1726776670.00057: variable 'omit' from source: magic vars 10472 1726776670.00063: starting attempt loop 10472 1726776670.00066: running the handler 10472 1726776670.00077: _low_level_execute_command(): starting 10472 1726776670.00084: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10472 1726776670.02387: stdout chunk (state=2): >>>/root <<< 10472 1726776670.02504: stderr chunk (state=3): >>><<< 10472 1726776670.02512: stdout chunk (state=3): >>><<< 10472 1726776670.02530: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10472 1726776670.02543: _low_level_execute_command(): starting 10472 1726776670.02549: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811 `" && echo ansible-tmp-1726776670.0253773-10472-31450016212811="` echo /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811 `" ) && sleep 0' 10472 1726776670.04988: stdout chunk (state=2): >>>ansible-tmp-1726776670.0253773-10472-31450016212811=/root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811 <<< 10472 1726776670.05118: stderr chunk (state=3): >>><<< 10472 1726776670.05124: stdout chunk (state=3): >>><<< 10472 1726776670.05139: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776670.0253773-10472-31450016212811=/root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811 , stderr= 10472 1726776670.05170: variable 'ansible_module_compression' from source: unknown 10472 1726776670.05208: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10472 1726776670.05234: variable 'ansible_facts' from source: unknown 10472 1726776670.05301: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/AnsiballZ_file.py 10472 1726776670.05394: Sending initial data 10472 1726776670.05401: Sent initial data (151 bytes) 10472 1726776670.07861: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpn69yyoed /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/AnsiballZ_file.py <<< 10472 1726776670.08846: stderr chunk (state=3): >>><<< 10472 1726776670.08853: stdout chunk (state=3): >>><<< 10472 1726776670.08871: done transferring module to remote 10472 1726776670.08881: _low_level_execute_command(): starting 10472 1726776670.08886: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/ /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/AnsiballZ_file.py && sleep 0' 10472 1726776670.11256: stderr chunk (state=2): >>><<< 10472 1726776670.11262: stdout chunk (state=2): >>><<< 10472 1726776670.11274: _low_level_execute_command() done: rc=0, stdout=, stderr= 10472 1726776670.11278: _low_level_execute_command(): starting 10472 1726776670.11283: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/AnsiballZ_file.py && sleep 0' 10472 1726776670.27502: stdout chunk (state=2): >>> {"path": "/etc/tuned/kernel_settings", "changed": false, "diff": {"before": {"path": "/etc/tuned/kernel_settings"}, "after": {"path": "/etc/tuned/kernel_settings"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0755", "state": "directory", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 24, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "directory", "mode": "0755", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10472 1726776670.28653: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10472 1726776670.28701: stderr chunk (state=3): >>><<< 10472 1726776670.28709: stdout chunk (state=3): >>><<< 10472 1726776670.28726: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/kernel_settings", "changed": false, "diff": {"before": {"path": "/etc/tuned/kernel_settings"}, "after": {"path": "/etc/tuned/kernel_settings"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0755", "state": "directory", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 24, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "directory", "mode": "0755", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10472 1726776670.28760: done with _execute_module (file, {'path': '/etc/tuned/kernel_settings', 'state': 'directory', 'mode': '0755', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10472 1726776670.28771: _low_level_execute_command(): starting 10472 1726776670.28777: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776670.0253773-10472-31450016212811/ > /dev/null 2>&1 && sleep 0' 10472 1726776670.31181: stderr chunk (state=2): >>><<< 10472 1726776670.31190: stdout chunk (state=2): >>><<< 10472 1726776670.31204: _low_level_execute_command() done: rc=0, stdout=, stderr= 10472 1726776670.31211: handler run complete 10472 1726776670.31232: attempt loop complete, returning result 10472 1726776670.31236: _execute() done 10472 1726776670.31240: dumping result to json 10472 1726776670.31245: done dumping result, returning 10472 1726776670.31253: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists [120fa90a-8a95-fee7-0cea-000000000409] 10472 1726776670.31260: sending task result for task 120fa90a-8a95-fee7-0cea-000000000409 10472 1726776670.31292: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000409 10472 1726776670.31296: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/tuned/kernel_settings", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 24, "state": "directory", "uid": 0 } 9704 1726776670.31454: no more pending results, returning what we have 9704 1726776670.31457: results queue empty 9704 1726776670.31458: checking for any_errors_fatal 9704 1726776670.31473: done checking for any_errors_fatal 9704 1726776670.31474: checking for max_fail_percentage 9704 1726776670.31475: done checking for max_fail_percentage 9704 1726776670.31476: checking to see if all hosts have failed and the running result is not ok 9704 1726776670.31476: done checking to see if all hosts have failed 9704 1726776670.31477: getting the remaining hosts for this loop 9704 1726776670.31478: done getting the remaining hosts for this loop 9704 1726776670.31482: getting the next task for host managed_node3 9704 1726776670.31487: done getting next task for host managed_node3 9704 1726776670.31490: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get active_profile 9704 1726776670.31493: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776670.31503: getting variables 9704 1726776670.31504: in VariableManager get_vars() 9704 1726776670.31538: Calling all_inventory to load vars for managed_node3 9704 1726776670.31543: Calling groups_inventory to load vars for managed_node3 9704 1726776670.31545: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776670.31554: Calling all_plugins_play to load vars for managed_node3 9704 1726776670.31556: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776670.31558: Calling groups_plugins_play to load vars for managed_node3 9704 1726776670.31665: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776670.31787: done with get_vars() 9704 1726776670.31794: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Get active_profile] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:80 Thursday 19 September 2024 16:11:10 -0400 (0:00:00.332) 0:00:23.247 **** 9704 1726776670.31864: entering _queue_task() for managed_node3/slurp 9704 1726776670.32030: worker is 1 (out of 1 available) 9704 1726776670.32046: exiting _queue_task() for managed_node3/slurp 9704 1726776670.32058: done queuing things up, now waiting for results queue to drain 9704 1726776670.32060: waiting for pending results... 10480 1726776670.32179: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get active_profile 10480 1726776670.32289: in run() - task 120fa90a-8a95-fee7-0cea-00000000040a 10480 1726776670.32304: variable 'ansible_search_path' from source: unknown 10480 1726776670.32308: variable 'ansible_search_path' from source: unknown 10480 1726776670.32336: calling self._execute() 10480 1726776670.32405: variable 'ansible_host' from source: host vars for 'managed_node3' 10480 1726776670.32414: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10480 1726776670.32422: variable 'omit' from source: magic vars 10480 1726776670.32498: variable 'omit' from source: magic vars 10480 1726776670.32532: variable 'omit' from source: magic vars 10480 1726776670.32553: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10480 1726776670.32764: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10480 1726776670.32823: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10480 1726776670.32855: variable 'omit' from source: magic vars 10480 1726776670.32886: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10480 1726776670.32913: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10480 1726776670.32933: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10480 1726776670.32948: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10480 1726776670.32960: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10480 1726776670.32983: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10480 1726776670.32988: variable 'ansible_host' from source: host vars for 'managed_node3' 10480 1726776670.32992: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10480 1726776670.33061: Set connection var ansible_module_compression to ZIP_DEFLATED 10480 1726776670.33069: Set connection var ansible_shell_type to sh 10480 1726776670.33077: Set connection var ansible_timeout to 10 10480 1726776670.33082: Set connection var ansible_pipelining to False 10480 1726776670.33089: Set connection var ansible_shell_executable to /bin/sh 10480 1726776670.33092: Set connection var ansible_connection to ssh 10480 1726776670.33109: variable 'ansible_shell_executable' from source: unknown 10480 1726776670.33114: variable 'ansible_connection' from source: unknown 10480 1726776670.33117: variable 'ansible_module_compression' from source: unknown 10480 1726776670.33121: variable 'ansible_shell_type' from source: unknown 10480 1726776670.33124: variable 'ansible_shell_executable' from source: unknown 10480 1726776670.33127: variable 'ansible_host' from source: host vars for 'managed_node3' 10480 1726776670.33133: variable 'ansible_pipelining' from source: unknown 10480 1726776670.33136: variable 'ansible_timeout' from source: unknown 10480 1726776670.33141: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10480 1726776670.33273: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10480 1726776670.33284: variable 'omit' from source: magic vars 10480 1726776670.33290: starting attempt loop 10480 1726776670.33293: running the handler 10480 1726776670.33304: _low_level_execute_command(): starting 10480 1726776670.33312: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10480 1726776670.35640: stdout chunk (state=2): >>>/root <<< 10480 1726776670.35757: stderr chunk (state=3): >>><<< 10480 1726776670.35763: stdout chunk (state=3): >>><<< 10480 1726776670.35780: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10480 1726776670.35794: _low_level_execute_command(): starting 10480 1726776670.35802: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371 `" && echo ansible-tmp-1726776670.3578823-10480-180624047636371="` echo /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371 `" ) && sleep 0' 10480 1726776670.38246: stdout chunk (state=2): >>>ansible-tmp-1726776670.3578823-10480-180624047636371=/root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371 <<< 10480 1726776670.38378: stderr chunk (state=3): >>><<< 10480 1726776670.38384: stdout chunk (state=3): >>><<< 10480 1726776670.38398: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776670.3578823-10480-180624047636371=/root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371 , stderr= 10480 1726776670.38434: variable 'ansible_module_compression' from source: unknown 10480 1726776670.38466: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.slurp-ZIP_DEFLATED 10480 1726776670.38492: variable 'ansible_facts' from source: unknown 10480 1726776670.38560: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/AnsiballZ_slurp.py 10480 1726776670.38738: Sending initial data 10480 1726776670.38747: Sent initial data (153 bytes) 10480 1726776670.41171: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp25y3e6nf /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/AnsiballZ_slurp.py <<< 10480 1726776670.42120: stderr chunk (state=3): >>><<< 10480 1726776670.42130: stdout chunk (state=3): >>><<< 10480 1726776670.42150: done transferring module to remote 10480 1726776670.42161: _low_level_execute_command(): starting 10480 1726776670.42167: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/ /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/AnsiballZ_slurp.py && sleep 0' 10480 1726776670.44507: stderr chunk (state=2): >>><<< 10480 1726776670.44515: stdout chunk (state=2): >>><<< 10480 1726776670.44533: _low_level_execute_command() done: rc=0, stdout=, stderr= 10480 1726776670.44538: _low_level_execute_command(): starting 10480 1726776670.44543: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/AnsiballZ_slurp.py && sleep 0' 10480 1726776670.59747: stdout chunk (state=2): >>> {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} <<< 10480 1726776670.60842: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10480 1726776670.60890: stderr chunk (state=3): >>><<< 10480 1726776670.60899: stdout chunk (state=3): >>><<< 10480 1726776670.60917: _low_level_execute_command() done: rc=0, stdout= {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} , stderr=Shared connection to 10.31.8.186 closed. 10480 1726776670.60948: done with _execute_module (slurp, {'path': '/etc/tuned/active_profile', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10480 1726776670.60961: _low_level_execute_command(): starting 10480 1726776670.60970: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776670.3578823-10480-180624047636371/ > /dev/null 2>&1 && sleep 0' 10480 1726776670.63622: stderr chunk (state=2): >>><<< 10480 1726776670.63635: stdout chunk (state=2): >>><<< 10480 1726776670.63653: _low_level_execute_command() done: rc=0, stdout=, stderr= 10480 1726776670.63661: handler run complete 10480 1726776670.63677: attempt loop complete, returning result 10480 1726776670.63682: _execute() done 10480 1726776670.63686: dumping result to json 10480 1726776670.63690: done dumping result, returning 10480 1726776670.63698: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get active_profile [120fa90a-8a95-fee7-0cea-00000000040a] 10480 1726776670.63705: sending task result for task 120fa90a-8a95-fee7-0cea-00000000040a 10480 1726776670.63742: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000040a 10480 1726776670.63747: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "encoding": "base64", "source": "/etc/tuned/active_profile" } 9704 1726776670.64140: no more pending results, returning what we have 9704 1726776670.64144: results queue empty 9704 1726776670.64145: checking for any_errors_fatal 9704 1726776670.64154: done checking for any_errors_fatal 9704 1726776670.64154: checking for max_fail_percentage 9704 1726776670.64156: done checking for max_fail_percentage 9704 1726776670.64156: checking to see if all hosts have failed and the running result is not ok 9704 1726776670.64157: done checking to see if all hosts have failed 9704 1726776670.64158: getting the remaining hosts for this loop 9704 1726776670.64159: done getting the remaining hosts for this loop 9704 1726776670.64162: getting the next task for host managed_node3 9704 1726776670.64167: done getting next task for host managed_node3 9704 1726776670.64170: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set active_profile 9704 1726776670.64172: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776670.64182: getting variables 9704 1726776670.64183: in VariableManager get_vars() 9704 1726776670.64214: Calling all_inventory to load vars for managed_node3 9704 1726776670.64217: Calling groups_inventory to load vars for managed_node3 9704 1726776670.64219: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776670.64227: Calling all_plugins_play to load vars for managed_node3 9704 1726776670.64230: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776670.64233: Calling groups_plugins_play to load vars for managed_node3 9704 1726776670.64383: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776670.64575: done with get_vars() 9704 1726776670.64585: done getting variables 9704 1726776670.64642: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set active_profile] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:85 Thursday 19 September 2024 16:11:10 -0400 (0:00:00.328) 0:00:23.575 **** 9704 1726776670.64670: entering _queue_task() for managed_node3/set_fact 9704 1726776670.64877: worker is 1 (out of 1 available) 9704 1726776670.64893: exiting _queue_task() for managed_node3/set_fact 9704 1726776670.64904: done queuing things up, now waiting for results queue to drain 9704 1726776670.64906: waiting for pending results... 10492 1726776670.65146: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set active_profile 10492 1726776670.65296: in run() - task 120fa90a-8a95-fee7-0cea-00000000040b 10492 1726776670.65313: variable 'ansible_search_path' from source: unknown 10492 1726776670.65318: variable 'ansible_search_path' from source: unknown 10492 1726776670.65351: calling self._execute() 10492 1726776670.65535: variable 'ansible_host' from source: host vars for 'managed_node3' 10492 1726776670.65544: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10492 1726776670.65553: variable 'omit' from source: magic vars 10492 1726776670.65653: variable 'omit' from source: magic vars 10492 1726776670.65701: variable 'omit' from source: magic vars 10492 1726776670.66114: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10492 1726776670.66125: variable '__cur_profile' from source: task vars 10492 1726776670.66270: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10492 1726776670.68677: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10492 1726776670.68742: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10492 1726776670.68794: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10492 1726776670.68828: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10492 1726776670.68857: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10492 1726776670.68930: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10492 1726776670.68959: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10492 1726776670.68984: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10492 1726776670.69025: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10492 1726776670.69042: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10492 1726776670.69152: variable '__kernel_settings_tuned_current_profile' from source: set_fact 10492 1726776670.69209: variable 'omit' from source: magic vars 10492 1726776670.69238: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10492 1726776670.69264: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10492 1726776670.69281: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10492 1726776670.69297: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10492 1726776670.69306: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10492 1726776670.69354: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10492 1726776670.69360: variable 'ansible_host' from source: host vars for 'managed_node3' 10492 1726776670.69364: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10492 1726776670.69455: Set connection var ansible_module_compression to ZIP_DEFLATED 10492 1726776670.69464: Set connection var ansible_shell_type to sh 10492 1726776670.69472: Set connection var ansible_timeout to 10 10492 1726776670.69477: Set connection var ansible_pipelining to False 10492 1726776670.69484: Set connection var ansible_shell_executable to /bin/sh 10492 1726776670.69486: Set connection var ansible_connection to ssh 10492 1726776670.69506: variable 'ansible_shell_executable' from source: unknown 10492 1726776670.69510: variable 'ansible_connection' from source: unknown 10492 1726776670.69513: variable 'ansible_module_compression' from source: unknown 10492 1726776670.69516: variable 'ansible_shell_type' from source: unknown 10492 1726776670.69518: variable 'ansible_shell_executable' from source: unknown 10492 1726776670.69521: variable 'ansible_host' from source: host vars for 'managed_node3' 10492 1726776670.69524: variable 'ansible_pipelining' from source: unknown 10492 1726776670.69527: variable 'ansible_timeout' from source: unknown 10492 1726776670.69532: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10492 1726776670.69619: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10492 1726776670.69633: variable 'omit' from source: magic vars 10492 1726776670.69640: starting attempt loop 10492 1726776670.69644: running the handler 10492 1726776670.69654: handler run complete 10492 1726776670.69664: attempt loop complete, returning result 10492 1726776670.69667: _execute() done 10492 1726776670.69670: dumping result to json 10492 1726776670.69673: done dumping result, returning 10492 1726776670.69679: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set active_profile [120fa90a-8a95-fee7-0cea-00000000040b] 10492 1726776670.69684: sending task result for task 120fa90a-8a95-fee7-0cea-00000000040b 10492 1726776670.69710: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000040b 10492 1726776670.69714: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_active_profile": "virtual-guest kernel_settings" }, "changed": false } 9704 1726776670.69978: no more pending results, returning what we have 9704 1726776670.69981: results queue empty 9704 1726776670.69982: checking for any_errors_fatal 9704 1726776670.69986: done checking for any_errors_fatal 9704 1726776670.69987: checking for max_fail_percentage 9704 1726776670.69988: done checking for max_fail_percentage 9704 1726776670.69989: checking to see if all hosts have failed and the running result is not ok 9704 1726776670.69989: done checking to see if all hosts have failed 9704 1726776670.69990: getting the remaining hosts for this loop 9704 1726776670.69991: done getting the remaining hosts for this loop 9704 1726776670.69994: getting the next task for host managed_node3 9704 1726776670.69999: done getting next task for host managed_node3 9704 1726776670.70002: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile 9704 1726776670.70008: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776670.70021: getting variables 9704 1726776670.70022: in VariableManager get_vars() 9704 1726776670.70048: Calling all_inventory to load vars for managed_node3 9704 1726776670.70050: Calling groups_inventory to load vars for managed_node3 9704 1726776670.70052: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776670.70058: Calling all_plugins_play to load vars for managed_node3 9704 1726776670.70060: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776670.70062: Calling groups_plugins_play to load vars for managed_node3 9704 1726776670.70162: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776670.70280: done with get_vars() 9704 1726776670.70287: done getting variables 9704 1726776670.70325: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:91 Thursday 19 September 2024 16:11:10 -0400 (0:00:00.056) 0:00:23.632 **** 9704 1726776670.70351: entering _queue_task() for managed_node3/copy 9704 1726776670.70511: worker is 1 (out of 1 available) 9704 1726776670.70527: exiting _queue_task() for managed_node3/copy 9704 1726776670.70540: done queuing things up, now waiting for results queue to drain 9704 1726776670.70545: waiting for pending results... 10495 1726776670.70664: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile 10495 1726776670.70779: in run() - task 120fa90a-8a95-fee7-0cea-00000000040c 10495 1726776670.70796: variable 'ansible_search_path' from source: unknown 10495 1726776670.70800: variable 'ansible_search_path' from source: unknown 10495 1726776670.70826: calling self._execute() 10495 1726776670.70892: variable 'ansible_host' from source: host vars for 'managed_node3' 10495 1726776670.70900: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10495 1726776670.70909: variable 'omit' from source: magic vars 10495 1726776670.70978: variable 'omit' from source: magic vars 10495 1726776670.71011: variable 'omit' from source: magic vars 10495 1726776670.71031: variable '__kernel_settings_active_profile' from source: set_fact 10495 1726776670.71239: variable '__kernel_settings_active_profile' from source: set_fact 10495 1726776670.71258: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10495 1726776670.71306: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10495 1726776670.71360: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10495 1726776670.71381: variable 'omit' from source: magic vars 10495 1726776670.71408: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10495 1726776670.71435: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10495 1726776670.71455: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10495 1726776670.71469: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10495 1726776670.71479: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10495 1726776670.71501: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10495 1726776670.71506: variable 'ansible_host' from source: host vars for 'managed_node3' 10495 1726776670.71510: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10495 1726776670.71578: Set connection var ansible_module_compression to ZIP_DEFLATED 10495 1726776670.71586: Set connection var ansible_shell_type to sh 10495 1726776670.71594: Set connection var ansible_timeout to 10 10495 1726776670.71599: Set connection var ansible_pipelining to False 10495 1726776670.71606: Set connection var ansible_shell_executable to /bin/sh 10495 1726776670.71609: Set connection var ansible_connection to ssh 10495 1726776670.71623: variable 'ansible_shell_executable' from source: unknown 10495 1726776670.71627: variable 'ansible_connection' from source: unknown 10495 1726776670.71632: variable 'ansible_module_compression' from source: unknown 10495 1726776670.71636: variable 'ansible_shell_type' from source: unknown 10495 1726776670.71638: variable 'ansible_shell_executable' from source: unknown 10495 1726776670.71639: variable 'ansible_host' from source: host vars for 'managed_node3' 10495 1726776670.71642: variable 'ansible_pipelining' from source: unknown 10495 1726776670.71644: variable 'ansible_timeout' from source: unknown 10495 1726776670.71647: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10495 1726776670.71725: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10495 1726776670.71738: variable 'omit' from source: magic vars 10495 1726776670.71745: starting attempt loop 10495 1726776670.71749: running the handler 10495 1726776670.71760: _low_level_execute_command(): starting 10495 1726776670.71768: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10495 1726776670.74881: stdout chunk (state=2): >>>/root <<< 10495 1726776670.75013: stderr chunk (state=3): >>><<< 10495 1726776670.75021: stdout chunk (state=3): >>><<< 10495 1726776670.75047: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10495 1726776670.75064: _low_level_execute_command(): starting 10495 1726776670.75071: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432 `" && echo ansible-tmp-1726776670.750579-10495-134697228432432="` echo /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432 `" ) && sleep 0' 10495 1726776670.78031: stdout chunk (state=2): >>>ansible-tmp-1726776670.750579-10495-134697228432432=/root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432 <<< 10495 1726776670.78154: stderr chunk (state=3): >>><<< 10495 1726776670.78162: stdout chunk (state=3): >>><<< 10495 1726776670.78177: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776670.750579-10495-134697228432432=/root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432 , stderr= 10495 1726776670.78248: variable 'ansible_module_compression' from source: unknown 10495 1726776670.78292: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10495 1726776670.78319: variable 'ansible_facts' from source: unknown 10495 1726776670.78391: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_stat.py 10495 1726776670.78504: Sending initial data 10495 1726776670.78512: Sent initial data (151 bytes) 10495 1726776670.81152: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpsyt8b5y_ /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_stat.py <<< 10495 1726776670.82110: stderr chunk (state=3): >>><<< 10495 1726776670.82118: stdout chunk (state=3): >>><<< 10495 1726776670.82138: done transferring module to remote 10495 1726776670.82151: _low_level_execute_command(): starting 10495 1726776670.82157: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/ /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_stat.py && sleep 0' 10495 1726776670.84626: stderr chunk (state=2): >>><<< 10495 1726776670.84635: stdout chunk (state=2): >>><<< 10495 1726776670.84648: _low_level_execute_command() done: rc=0, stdout=, stderr= 10495 1726776670.84652: _low_level_execute_command(): starting 10495 1726776670.84657: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_stat.py && sleep 0' 10495 1726776671.00746: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 30, "inode": 142606531, "dev": 51713, "nlink": 1, "atime": 1726776663.9393117, "mtime": 1726776658.1132364, "ctime": 1726776658.1132364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "mimetype": "text/plain", "charset": "us-ascii", "version": "3590528432", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10495 1726776671.01895: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10495 1726776671.01941: stderr chunk (state=3): >>><<< 10495 1726776671.01948: stdout chunk (state=3): >>><<< 10495 1726776671.01963: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 30, "inode": 142606531, "dev": 51713, "nlink": 1, "atime": 1726776663.9393117, "mtime": 1726776658.1132364, "ctime": 1726776658.1132364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "mimetype": "text/plain", "charset": "us-ascii", "version": "3590528432", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10495 1726776671.02003: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/active_profile', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10495 1726776671.02039: variable 'ansible_module_compression' from source: unknown 10495 1726776671.02071: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10495 1726776671.02090: variable 'ansible_facts' from source: unknown 10495 1726776671.02150: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_file.py 10495 1726776671.02235: Sending initial data 10495 1726776671.02243: Sent initial data (151 bytes) 10495 1726776671.04901: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmptg6q49cb /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_file.py <<< 10495 1726776671.06529: stderr chunk (state=3): >>><<< 10495 1726776671.06537: stdout chunk (state=3): >>><<< 10495 1726776671.06557: done transferring module to remote 10495 1726776671.06566: _low_level_execute_command(): starting 10495 1726776671.06570: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/ /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_file.py && sleep 0' 10495 1726776671.09206: stderr chunk (state=2): >>><<< 10495 1726776671.09213: stdout chunk (state=2): >>><<< 10495 1726776671.09226: _low_level_execute_command() done: rc=0, stdout=, stderr= 10495 1726776671.09234: _low_level_execute_command(): starting 10495 1726776671.09239: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/AnsiballZ_file.py && sleep 0' 10495 1726776671.25237: stdout chunk (state=2): >>> {"path": "/etc/tuned/active_profile", "changed": false, "diff": {"before": {"path": "/etc/tuned/active_profile"}, "after": {"path": "/etc/tuned/active_profile"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/active_profile", "_original_basename": "tmptz_53ust", "recurse": false, "state": "file", "path": "/etc/tuned/active_profile", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10495 1726776671.26397: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10495 1726776671.26449: stderr chunk (state=3): >>><<< 10495 1726776671.26456: stdout chunk (state=3): >>><<< 10495 1726776671.26472: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/active_profile", "changed": false, "diff": {"before": {"path": "/etc/tuned/active_profile"}, "after": {"path": "/etc/tuned/active_profile"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/active_profile", "_original_basename": "tmptz_53ust", "recurse": false, "state": "file", "path": "/etc/tuned/active_profile", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10495 1726776671.26500: done with _execute_module (ansible.legacy.file, {'mode': '0600', 'dest': '/etc/tuned/active_profile', '_original_basename': 'tmptz_53ust', 'recurse': False, 'state': 'file', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10495 1726776671.26512: _low_level_execute_command(): starting 10495 1726776671.26518: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776670.750579-10495-134697228432432/ > /dev/null 2>&1 && sleep 0' 10495 1726776671.28905: stderr chunk (state=2): >>><<< 10495 1726776671.28912: stdout chunk (state=2): >>><<< 10495 1726776671.28925: _low_level_execute_command() done: rc=0, stdout=, stderr= 10495 1726776671.28935: handler run complete 10495 1726776671.28957: attempt loop complete, returning result 10495 1726776671.28964: _execute() done 10495 1726776671.28967: dumping result to json 10495 1726776671.28973: done dumping result, returning 10495 1726776671.28980: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile [120fa90a-8a95-fee7-0cea-00000000040c] 10495 1726776671.28986: sending task result for task 120fa90a-8a95-fee7-0cea-00000000040c 10495 1726776671.29020: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000040c 10495 1726776671.29024: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "dest": "/etc/tuned/active_profile", "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/etc/tuned/active_profile", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "state": "file", "uid": 0 } 9704 1726776671.29184: no more pending results, returning what we have 9704 1726776671.29187: results queue empty 9704 1726776671.29188: checking for any_errors_fatal 9704 1726776671.29195: done checking for any_errors_fatal 9704 1726776671.29195: checking for max_fail_percentage 9704 1726776671.29197: done checking for max_fail_percentage 9704 1726776671.29197: checking to see if all hosts have failed and the running result is not ok 9704 1726776671.29198: done checking to see if all hosts have failed 9704 1726776671.29198: getting the remaining hosts for this loop 9704 1726776671.29200: done getting the remaining hosts for this loop 9704 1726776671.29203: getting the next task for host managed_node3 9704 1726776671.29209: done getting next task for host managed_node3 9704 1726776671.29212: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual 9704 1726776671.29214: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776671.29224: getting variables 9704 1726776671.29225: in VariableManager get_vars() 9704 1726776671.29260: Calling all_inventory to load vars for managed_node3 9704 1726776671.29263: Calling groups_inventory to load vars for managed_node3 9704 1726776671.29265: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776671.29273: Calling all_plugins_play to load vars for managed_node3 9704 1726776671.29276: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776671.29278: Calling groups_plugins_play to load vars for managed_node3 9704 1726776671.29388: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776671.29686: done with get_vars() 9704 1726776671.29694: done getting variables 9704 1726776671.29736: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set profile_mode to manual] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:99 Thursday 19 September 2024 16:11:11 -0400 (0:00:00.594) 0:00:24.226 **** 9704 1726776671.29757: entering _queue_task() for managed_node3/copy 9704 1726776671.29921: worker is 1 (out of 1 available) 9704 1726776671.29937: exiting _queue_task() for managed_node3/copy 9704 1726776671.29949: done queuing things up, now waiting for results queue to drain 9704 1726776671.29951: waiting for pending results... 10523 1726776671.30078: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual 10523 1726776671.30195: in run() - task 120fa90a-8a95-fee7-0cea-00000000040d 10523 1726776671.30212: variable 'ansible_search_path' from source: unknown 10523 1726776671.30216: variable 'ansible_search_path' from source: unknown 10523 1726776671.30248: calling self._execute() 10523 1726776671.30317: variable 'ansible_host' from source: host vars for 'managed_node3' 10523 1726776671.30325: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10523 1726776671.30335: variable 'omit' from source: magic vars 10523 1726776671.30410: variable 'omit' from source: magic vars 10523 1726776671.30447: variable 'omit' from source: magic vars 10523 1726776671.30467: variable '__kernel_settings_tuned_profile_mode' from source: role '' all vars 10523 1726776671.30689: variable '__kernel_settings_tuned_profile_mode' from source: role '' all vars 10523 1726776671.30753: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10523 1726776671.30780: variable 'omit' from source: magic vars 10523 1726776671.30809: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10523 1726776671.30838: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10523 1726776671.30861: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10523 1726776671.30876: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10523 1726776671.30886: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10523 1726776671.30910: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10523 1726776671.30915: variable 'ansible_host' from source: host vars for 'managed_node3' 10523 1726776671.30919: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10523 1726776671.30990: Set connection var ansible_module_compression to ZIP_DEFLATED 10523 1726776671.30998: Set connection var ansible_shell_type to sh 10523 1726776671.31007: Set connection var ansible_timeout to 10 10523 1726776671.31012: Set connection var ansible_pipelining to False 10523 1726776671.31019: Set connection var ansible_shell_executable to /bin/sh 10523 1726776671.31022: Set connection var ansible_connection to ssh 10523 1726776671.31043: variable 'ansible_shell_executable' from source: unknown 10523 1726776671.31048: variable 'ansible_connection' from source: unknown 10523 1726776671.31051: variable 'ansible_module_compression' from source: unknown 10523 1726776671.31054: variable 'ansible_shell_type' from source: unknown 10523 1726776671.31057: variable 'ansible_shell_executable' from source: unknown 10523 1726776671.31061: variable 'ansible_host' from source: host vars for 'managed_node3' 10523 1726776671.31065: variable 'ansible_pipelining' from source: unknown 10523 1726776671.31068: variable 'ansible_timeout' from source: unknown 10523 1726776671.31074: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10523 1726776671.31165: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10523 1726776671.31177: variable 'omit' from source: magic vars 10523 1726776671.31183: starting attempt loop 10523 1726776671.31187: running the handler 10523 1726776671.31198: _low_level_execute_command(): starting 10523 1726776671.31206: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10523 1726776671.33549: stdout chunk (state=2): >>>/root <<< 10523 1726776671.33673: stderr chunk (state=3): >>><<< 10523 1726776671.33680: stdout chunk (state=3): >>><<< 10523 1726776671.33698: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10523 1726776671.33711: _low_level_execute_command(): starting 10523 1726776671.33718: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202 `" && echo ansible-tmp-1726776671.3370519-10523-216824249579202="` echo /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202 `" ) && sleep 0' 10523 1726776671.36239: stdout chunk (state=2): >>>ansible-tmp-1726776671.3370519-10523-216824249579202=/root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202 <<< 10523 1726776671.36422: stderr chunk (state=3): >>><<< 10523 1726776671.36431: stdout chunk (state=3): >>><<< 10523 1726776671.36451: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776671.3370519-10523-216824249579202=/root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202 , stderr= 10523 1726776671.36540: variable 'ansible_module_compression' from source: unknown 10523 1726776671.36603: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10523 1726776671.36639: variable 'ansible_facts' from source: unknown 10523 1726776671.36748: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_stat.py 10523 1726776671.36905: Sending initial data 10523 1726776671.36913: Sent initial data (152 bytes) 10523 1726776671.39916: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpm9uqcg87 /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_stat.py <<< 10523 1726776671.40608: stderr chunk (state=3): >>><<< 10523 1726776671.40616: stdout chunk (state=3): >>><<< 10523 1726776671.40642: done transferring module to remote 10523 1726776671.40654: _low_level_execute_command(): starting 10523 1726776671.40660: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/ /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_stat.py && sleep 0' 10523 1726776671.43029: stderr chunk (state=2): >>><<< 10523 1726776671.43040: stdout chunk (state=2): >>><<< 10523 1726776671.43056: _low_level_execute_command() done: rc=0, stdout=, stderr= 10523 1726776671.43061: _low_level_execute_command(): starting 10523 1726776671.43066: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_stat.py && sleep 0' 10523 1726776671.59103: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 7, "inode": 148897989, "dev": 51713, "nlink": 1, "atime": 1726776664.3133166, "mtime": 1726776658.1142364, "ctime": 1726776658.1142364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "mimetype": "text/plain", "charset": "us-ascii", "version": "1846195820", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10523 1726776671.60288: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10523 1726776671.60337: stderr chunk (state=3): >>><<< 10523 1726776671.60344: stdout chunk (state=3): >>><<< 10523 1726776671.60361: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 7, "inode": 148897989, "dev": 51713, "nlink": 1, "atime": 1726776664.3133166, "mtime": 1726776658.1142364, "ctime": 1726776658.1142364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "mimetype": "text/plain", "charset": "us-ascii", "version": "1846195820", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10523 1726776671.60402: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/profile_mode', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10523 1726776671.60442: variable 'ansible_module_compression' from source: unknown 10523 1726776671.60475: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10523 1726776671.60495: variable 'ansible_facts' from source: unknown 10523 1726776671.60554: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_file.py 10523 1726776671.60644: Sending initial data 10523 1726776671.60651: Sent initial data (152 bytes) 10523 1726776671.63162: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpmh8q3uug /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_file.py <<< 10523 1726776671.64160: stderr chunk (state=3): >>><<< 10523 1726776671.64166: stdout chunk (state=3): >>><<< 10523 1726776671.64183: done transferring module to remote 10523 1726776671.64192: _low_level_execute_command(): starting 10523 1726776671.64197: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/ /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_file.py && sleep 0' 10523 1726776671.66550: stderr chunk (state=2): >>><<< 10523 1726776671.66557: stdout chunk (state=2): >>><<< 10523 1726776671.66570: _low_level_execute_command() done: rc=0, stdout=, stderr= 10523 1726776671.66575: _low_level_execute_command(): starting 10523 1726776671.66582: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/AnsiballZ_file.py && sleep 0' 10523 1726776671.82817: stdout chunk (state=2): >>> {"path": "/etc/tuned/profile_mode", "changed": false, "diff": {"before": {"path": "/etc/tuned/profile_mode"}, "after": {"path": "/etc/tuned/profile_mode"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/profile_mode", "_original_basename": "tmpuh607qxp", "recurse": false, "state": "file", "path": "/etc/tuned/profile_mode", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10523 1726776671.83978: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10523 1726776671.84019: stderr chunk (state=3): >>><<< 10523 1726776671.84026: stdout chunk (state=3): >>><<< 10523 1726776671.84045: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/profile_mode", "changed": false, "diff": {"before": {"path": "/etc/tuned/profile_mode"}, "after": {"path": "/etc/tuned/profile_mode"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/profile_mode", "_original_basename": "tmpuh607qxp", "recurse": false, "state": "file", "path": "/etc/tuned/profile_mode", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10523 1726776671.84074: done with _execute_module (ansible.legacy.file, {'mode': '0600', 'dest': '/etc/tuned/profile_mode', '_original_basename': 'tmpuh607qxp', 'recurse': False, 'state': 'file', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10523 1726776671.84085: _low_level_execute_command(): starting 10523 1726776671.84090: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776671.3370519-10523-216824249579202/ > /dev/null 2>&1 && sleep 0' 10523 1726776671.86462: stderr chunk (state=2): >>><<< 10523 1726776671.86469: stdout chunk (state=2): >>><<< 10523 1726776671.86482: _low_level_execute_command() done: rc=0, stdout=, stderr= 10523 1726776671.86490: handler run complete 10523 1726776671.86510: attempt loop complete, returning result 10523 1726776671.86514: _execute() done 10523 1726776671.86518: dumping result to json 10523 1726776671.86523: done dumping result, returning 10523 1726776671.86532: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual [120fa90a-8a95-fee7-0cea-00000000040d] 10523 1726776671.86538: sending task result for task 120fa90a-8a95-fee7-0cea-00000000040d 10523 1726776671.86571: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000040d 10523 1726776671.86575: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "dest": "/etc/tuned/profile_mode", "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/etc/tuned/profile_mode", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "state": "file", "uid": 0 } 9704 1726776671.86735: no more pending results, returning what we have 9704 1726776671.86738: results queue empty 9704 1726776671.86739: checking for any_errors_fatal 9704 1726776671.86748: done checking for any_errors_fatal 9704 1726776671.86748: checking for max_fail_percentage 9704 1726776671.86749: done checking for max_fail_percentage 9704 1726776671.86750: checking to see if all hosts have failed and the running result is not ok 9704 1726776671.86751: done checking to see if all hosts have failed 9704 1726776671.86752: getting the remaining hosts for this loop 9704 1726776671.86753: done getting the remaining hosts for this loop 9704 1726776671.86756: getting the next task for host managed_node3 9704 1726776671.86762: done getting next task for host managed_node3 9704 1726776671.86765: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get current config 9704 1726776671.86767: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776671.86778: getting variables 9704 1726776671.86779: in VariableManager get_vars() 9704 1726776671.86810: Calling all_inventory to load vars for managed_node3 9704 1726776671.86813: Calling groups_inventory to load vars for managed_node3 9704 1726776671.86815: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776671.86823: Calling all_plugins_play to load vars for managed_node3 9704 1726776671.86825: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776671.86827: Calling groups_plugins_play to load vars for managed_node3 9704 1726776671.86938: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776671.87073: done with get_vars() 9704 1726776671.87081: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Get current config] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:107 Thursday 19 September 2024 16:11:11 -0400 (0:00:00.573) 0:00:24.800 **** 9704 1726776671.87141: entering _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776671.87301: worker is 1 (out of 1 available) 9704 1726776671.87315: exiting _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776671.87326: done queuing things up, now waiting for results queue to drain 9704 1726776671.87330: waiting for pending results... 10552 1726776671.87454: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get current config 10552 1726776671.87569: in run() - task 120fa90a-8a95-fee7-0cea-00000000040e 10552 1726776671.87584: variable 'ansible_search_path' from source: unknown 10552 1726776671.87588: variable 'ansible_search_path' from source: unknown 10552 1726776671.87615: calling self._execute() 10552 1726776671.87684: variable 'ansible_host' from source: host vars for 'managed_node3' 10552 1726776671.87692: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10552 1726776671.87701: variable 'omit' from source: magic vars 10552 1726776671.87773: variable 'omit' from source: magic vars 10552 1726776671.87805: variable 'omit' from source: magic vars 10552 1726776671.87825: variable '__kernel_settings_profile_filename' from source: role '' all vars 10552 1726776671.88054: variable '__kernel_settings_profile_filename' from source: role '' all vars 10552 1726776671.88131: variable '__kernel_settings_profile_dir' from source: role '' all vars 10552 1726776671.88204: variable '__kernel_settings_profile_parent' from source: set_fact 10552 1726776671.88214: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10552 1726776671.88323: variable 'omit' from source: magic vars 10552 1726776671.88375: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10552 1726776671.88409: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10552 1726776671.88428: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10552 1726776671.88446: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10552 1726776671.88456: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10552 1726776671.88485: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10552 1726776671.88490: variable 'ansible_host' from source: host vars for 'managed_node3' 10552 1726776671.88492: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10552 1726776671.88581: Set connection var ansible_module_compression to ZIP_DEFLATED 10552 1726776671.88590: Set connection var ansible_shell_type to sh 10552 1726776671.88598: Set connection var ansible_timeout to 10 10552 1726776671.88604: Set connection var ansible_pipelining to False 10552 1726776671.88611: Set connection var ansible_shell_executable to /bin/sh 10552 1726776671.88614: Set connection var ansible_connection to ssh 10552 1726776671.88633: variable 'ansible_shell_executable' from source: unknown 10552 1726776671.88637: variable 'ansible_connection' from source: unknown 10552 1726776671.88640: variable 'ansible_module_compression' from source: unknown 10552 1726776671.88643: variable 'ansible_shell_type' from source: unknown 10552 1726776671.88645: variable 'ansible_shell_executable' from source: unknown 10552 1726776671.88648: variable 'ansible_host' from source: host vars for 'managed_node3' 10552 1726776671.88651: variable 'ansible_pipelining' from source: unknown 10552 1726776671.88654: variable 'ansible_timeout' from source: unknown 10552 1726776671.88657: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10552 1726776671.88815: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10552 1726776671.88826: variable 'omit' from source: magic vars 10552 1726776671.88833: starting attempt loop 10552 1726776671.88840: running the handler 10552 1726776671.88852: _low_level_execute_command(): starting 10552 1726776671.88860: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10552 1726776671.91374: stdout chunk (state=2): >>>/root <<< 10552 1726776671.91492: stderr chunk (state=3): >>><<< 10552 1726776671.91499: stdout chunk (state=3): >>><<< 10552 1726776671.91516: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10552 1726776671.91533: _low_level_execute_command(): starting 10552 1726776671.91539: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308 `" && echo ansible-tmp-1726776671.9152431-10552-158227425339308="` echo /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308 `" ) && sleep 0' 10552 1726776671.93959: stdout chunk (state=2): >>>ansible-tmp-1726776671.9152431-10552-158227425339308=/root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308 <<< 10552 1726776671.94082: stderr chunk (state=3): >>><<< 10552 1726776671.94088: stdout chunk (state=3): >>><<< 10552 1726776671.94102: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776671.9152431-10552-158227425339308=/root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308 , stderr= 10552 1726776671.94139: variable 'ansible_module_compression' from source: unknown 10552 1726776671.94169: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.kernel_settings_get_config-ZIP_DEFLATED 10552 1726776671.94200: variable 'ansible_facts' from source: unknown 10552 1726776671.94266: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/AnsiballZ_kernel_settings_get_config.py 10552 1726776671.94360: Sending initial data 10552 1726776671.94367: Sent initial data (174 bytes) 10552 1726776671.96821: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpe_6csx68 /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/AnsiballZ_kernel_settings_get_config.py <<< 10552 1726776671.97783: stderr chunk (state=3): >>><<< 10552 1726776671.97790: stdout chunk (state=3): >>><<< 10552 1726776671.97809: done transferring module to remote 10552 1726776671.97819: _low_level_execute_command(): starting 10552 1726776671.97824: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/ /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10552 1726776672.00177: stderr chunk (state=2): >>><<< 10552 1726776672.00183: stdout chunk (state=2): >>><<< 10552 1726776672.00196: _low_level_execute_command() done: rc=0, stdout=, stderr= 10552 1726776672.00201: _low_level_execute_command(): starting 10552 1726776672.00206: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10552 1726776672.15932: stdout chunk (state=2): >>> {"changed": false, "data": {"main": {"summary": "kernel settings"}, "sysctl": {"fs.epoll.max_user_watches": "785592", "fs.file-max": "379724"}, "sysfs": {"/sys/kernel/debug/x86/ibrs_enabled": "0", "/sys/kernel/debug/x86/pti_enabled": "0", "/sys/kernel/debug/x86/retp_enabled": "0"}, "vm": {"transparent_hugepages": "madvise"}}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}} <<< 10552 1726776672.17001: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10552 1726776672.17050: stderr chunk (state=3): >>><<< 10552 1726776672.17057: stdout chunk (state=3): >>><<< 10552 1726776672.17074: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "data": {"main": {"summary": "kernel settings"}, "sysctl": {"fs.epoll.max_user_watches": "785592", "fs.file-max": "379724"}, "sysfs": {"/sys/kernel/debug/x86/ibrs_enabled": "0", "/sys/kernel/debug/x86/pti_enabled": "0", "/sys/kernel/debug/x86/retp_enabled": "0"}, "vm": {"transparent_hugepages": "madvise"}}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 10552 1726776672.17099: done with _execute_module (fedora.linux_system_roles.kernel_settings_get_config, {'path': '/etc/tuned/kernel_settings/tuned.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'fedora.linux_system_roles.kernel_settings_get_config', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10552 1726776672.17110: _low_level_execute_command(): starting 10552 1726776672.17116: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776671.9152431-10552-158227425339308/ > /dev/null 2>&1 && sleep 0' 10552 1726776672.19501: stderr chunk (state=2): >>><<< 10552 1726776672.19508: stdout chunk (state=2): >>><<< 10552 1726776672.19521: _low_level_execute_command() done: rc=0, stdout=, stderr= 10552 1726776672.19527: handler run complete 10552 1726776672.19544: attempt loop complete, returning result 10552 1726776672.19548: _execute() done 10552 1726776672.19552: dumping result to json 10552 1726776672.19556: done dumping result, returning 10552 1726776672.19563: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get current config [120fa90a-8a95-fee7-0cea-00000000040e] 10552 1726776672.19570: sending task result for task 120fa90a-8a95-fee7-0cea-00000000040e 10552 1726776672.19600: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000040e 10552 1726776672.19604: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "data": { "main": { "summary": "kernel settings" }, "sysctl": { "fs.epoll.max_user_watches": "785592", "fs.file-max": "379724" }, "sysfs": { "/sys/kernel/debug/x86/ibrs_enabled": "0", "/sys/kernel/debug/x86/pti_enabled": "0", "/sys/kernel/debug/x86/retp_enabled": "0" }, "vm": { "transparent_hugepages": "madvise" } } } 9704 1726776672.19775: no more pending results, returning what we have 9704 1726776672.19779: results queue empty 9704 1726776672.19780: checking for any_errors_fatal 9704 1726776672.19786: done checking for any_errors_fatal 9704 1726776672.19787: checking for max_fail_percentage 9704 1726776672.19788: done checking for max_fail_percentage 9704 1726776672.19790: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.19791: done checking to see if all hosts have failed 9704 1726776672.19791: getting the remaining hosts for this loop 9704 1726776672.19792: done getting the remaining hosts for this loop 9704 1726776672.19795: getting the next task for host managed_node3 9704 1726776672.19802: done getting next task for host managed_node3 9704 1726776672.19805: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings 9704 1726776672.19807: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.19818: getting variables 9704 1726776672.19819: in VariableManager get_vars() 9704 1726776672.19856: Calling all_inventory to load vars for managed_node3 9704 1726776672.19859: Calling groups_inventory to load vars for managed_node3 9704 1726776672.19861: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.19868: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.19870: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.19872: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.20022: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.20148: done with get_vars() 9704 1726776672.20155: done getting variables 9704 1726776672.20197: Loading ActionModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/template.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Apply kernel settings] ******* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:112 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.330) 0:00:25.130 **** 9704 1726776672.20220: entering _queue_task() for managed_node3/template 9704 1726776672.20387: worker is 1 (out of 1 available) 9704 1726776672.20401: exiting _queue_task() for managed_node3/template 9704 1726776672.20412: done queuing things up, now waiting for results queue to drain 9704 1726776672.20414: waiting for pending results... 10562 1726776672.20535: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings 10562 1726776672.20646: in run() - task 120fa90a-8a95-fee7-0cea-00000000040f 10562 1726776672.20664: variable 'ansible_search_path' from source: unknown 10562 1726776672.20668: variable 'ansible_search_path' from source: unknown 10562 1726776672.20697: calling self._execute() 10562 1726776672.20767: variable 'ansible_host' from source: host vars for 'managed_node3' 10562 1726776672.20775: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10562 1726776672.20784: variable 'omit' from source: magic vars 10562 1726776672.20857: variable 'omit' from source: magic vars 10562 1726776672.20894: variable 'omit' from source: magic vars 10562 1726776672.21127: variable '__kernel_settings_profile_src' from source: role '' all vars 10562 1726776672.21138: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10562 1726776672.21195: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10562 1726776672.21215: variable '__kernel_settings_profile_filename' from source: role '' all vars 10562 1726776672.21261: variable '__kernel_settings_profile_filename' from source: role '' all vars 10562 1726776672.21314: variable '__kernel_settings_profile_dir' from source: role '' all vars 10562 1726776672.21375: variable '__kernel_settings_profile_parent' from source: set_fact 10562 1726776672.21389: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10562 1726776672.21426: variable 'omit' from source: magic vars 10562 1726776672.21475: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10562 1726776672.21502: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10562 1726776672.21517: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10562 1726776672.21530: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10562 1726776672.21541: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10562 1726776672.21566: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10562 1726776672.21571: variable 'ansible_host' from source: host vars for 'managed_node3' 10562 1726776672.21576: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10562 1726776672.21642: Set connection var ansible_module_compression to ZIP_DEFLATED 10562 1726776672.21650: Set connection var ansible_shell_type to sh 10562 1726776672.21658: Set connection var ansible_timeout to 10 10562 1726776672.21663: Set connection var ansible_pipelining to False 10562 1726776672.21669: Set connection var ansible_shell_executable to /bin/sh 10562 1726776672.21672: Set connection var ansible_connection to ssh 10562 1726776672.21685: variable 'ansible_shell_executable' from source: unknown 10562 1726776672.21688: variable 'ansible_connection' from source: unknown 10562 1726776672.21690: variable 'ansible_module_compression' from source: unknown 10562 1726776672.21691: variable 'ansible_shell_type' from source: unknown 10562 1726776672.21693: variable 'ansible_shell_executable' from source: unknown 10562 1726776672.21694: variable 'ansible_host' from source: host vars for 'managed_node3' 10562 1726776672.21696: variable 'ansible_pipelining' from source: unknown 10562 1726776672.21698: variable 'ansible_timeout' from source: unknown 10562 1726776672.21701: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10562 1726776672.21792: Loading ActionModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/template.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10562 1726776672.21804: variable 'omit' from source: magic vars 10562 1726776672.21810: starting attempt loop 10562 1726776672.21813: running the handler 10562 1726776672.21821: _low_level_execute_command(): starting 10562 1726776672.21827: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10562 1726776672.24115: stdout chunk (state=2): >>>/root <<< 10562 1726776672.24232: stderr chunk (state=3): >>><<< 10562 1726776672.24238: stdout chunk (state=3): >>><<< 10562 1726776672.24254: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10562 1726776672.24267: _low_level_execute_command(): starting 10562 1726776672.24274: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727 `" && echo ansible-tmp-1726776672.2426186-10562-26494737368727="` echo /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727 `" ) && sleep 0' 10562 1726776672.26673: stdout chunk (state=2): >>>ansible-tmp-1726776672.2426186-10562-26494737368727=/root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727 <<< 10562 1726776672.26886: stderr chunk (state=3): >>><<< 10562 1726776672.26893: stdout chunk (state=3): >>><<< 10562 1726776672.26909: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776672.2426186-10562-26494737368727=/root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727 , stderr= 10562 1726776672.26927: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks 10562 1726776672.26952: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/kernel_settings.j2 10562 1726776672.26974: variable 'ansible_search_path' from source: unknown 10562 1726776672.27885: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10562 1726776672.29951: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10562 1726776672.29999: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10562 1726776672.30032: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10562 1726776672.30061: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10562 1726776672.30095: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10562 1726776672.30279: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10562 1726776672.30303: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10562 1726776672.30326: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10562 1726776672.30356: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10562 1726776672.30369: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10562 1726776672.30590: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10562 1726776672.30609: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10562 1726776672.30630: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10562 1726776672.30659: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10562 1726776672.30670: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10562 1726776672.30920: variable 'ansible_managed' from source: unknown 10562 1726776672.30928: variable '__sections' from source: task vars 10562 1726776672.31015: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10562 1726776672.31035: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10562 1726776672.31055: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10562 1726776672.31081: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10562 1726776672.31092: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10562 1726776672.31172: variable 'kernel_settings_sysctl' from source: include_vars 10562 1726776672.31183: variable '__kernel_settings_state_empty' from source: role '' all vars 10562 1726776672.31189: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10562 1726776672.31242: variable '__sysctl_old' from source: task vars 10562 1726776672.31291: variable '__sysctl_old' from source: task vars 10562 1726776672.31431: variable 'kernel_settings_purge' from source: role '' defaults 10562 1726776672.31440: variable 'kernel_settings_sysctl' from source: include_vars 10562 1726776672.31451: variable '__kernel_settings_state_empty' from source: role '' all vars 10562 1726776672.31457: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10562 1726776672.31462: variable '__kernel_settings_profile_contents' from source: set_fact 10562 1726776672.31595: variable 'kernel_settings_sysfs' from source: include_vars 10562 1726776672.31606: variable '__kernel_settings_state_empty' from source: role '' all vars 10562 1726776672.31612: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10562 1726776672.31646: variable '__sysfs_old' from source: task vars 10562 1726776672.31700: variable '__sysfs_old' from source: task vars 10562 1726776672.31881: variable 'kernel_settings_purge' from source: role '' defaults 10562 1726776672.31888: variable 'kernel_settings_sysfs' from source: include_vars 10562 1726776672.31897: variable '__kernel_settings_state_empty' from source: role '' all vars 10562 1726776672.31901: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10562 1726776672.31906: variable '__kernel_settings_profile_contents' from source: set_fact 10562 1726776672.31936: variable 'kernel_settings_systemd_cpu_affinity' from source: include_vars 10562 1726776672.31978: variable 'kernel_settings_systemd_cpu_affinity' from source: include_vars 10562 1726776672.31992: variable '__systemd_old' from source: task vars 10562 1726776672.32034: variable '__systemd_old' from source: task vars 10562 1726776672.32160: variable 'kernel_settings_purge' from source: role '' defaults 10562 1726776672.32166: variable 'kernel_settings_systemd_cpu_affinity' from source: include_vars 10562 1726776672.32172: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32177: variable '__kernel_settings_profile_contents' from source: set_fact 10562 1726776672.32189: variable 'kernel_settings_transparent_hugepages' from source: include_vars 10562 1726776672.32231: variable 'kernel_settings_transparent_hugepages' from source: include_vars 10562 1726776672.32240: variable 'kernel_settings_transparent_hugepages_defrag' from source: include_vars 10562 1726776672.32281: variable 'kernel_settings_transparent_hugepages_defrag' from source: include_vars 10562 1726776672.32296: variable '__trans_huge_old' from source: task vars 10562 1726776672.32341: variable '__trans_huge_old' from source: task vars 10562 1726776672.32537: variable 'kernel_settings_purge' from source: role '' defaults 10562 1726776672.32544: variable 'kernel_settings_transparent_hugepages' from source: include_vars 10562 1726776672.32549: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32554: variable '__kernel_settings_profile_contents' from source: set_fact 10562 1726776672.32565: variable '__trans_defrag_old' from source: task vars 10562 1726776672.32620: variable '__trans_defrag_old' from source: task vars 10562 1726776672.32814: variable 'kernel_settings_purge' from source: role '' defaults 10562 1726776672.32821: variable 'kernel_settings_transparent_hugepages_defrag' from source: include_vars 10562 1726776672.32827: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32834: variable '__kernel_settings_profile_contents' from source: set_fact 10562 1726776672.32851: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32864: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32873: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32885: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32893: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32899: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32906: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32919: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32931: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32939: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32953: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32963: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32970: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32976: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32983: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32989: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.32995: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.33001: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.33007: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.33012: variable '__kernel_settings_state_absent' from source: role '' all vars 10562 1726776672.33475: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10562 1726776672.33515: variable 'ansible_module_compression' from source: unknown 10562 1726776672.33557: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10562 1726776672.33576: variable 'ansible_facts' from source: unknown 10562 1726776672.33641: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_stat.py 10562 1726776672.33725: Sending initial data 10562 1726776672.33733: Sent initial data (151 bytes) 10562 1726776672.36419: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpsw0u7qvh /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_stat.py <<< 10562 1726776672.37373: stderr chunk (state=3): >>><<< 10562 1726776672.37380: stdout chunk (state=3): >>><<< 10562 1726776672.37395: done transferring module to remote 10562 1726776672.37404: _low_level_execute_command(): starting 10562 1726776672.37408: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/ /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_stat.py && sleep 0' 10562 1726776672.40009: stderr chunk (state=2): >>><<< 10562 1726776672.40017: stdout chunk (state=2): >>><<< 10562 1726776672.40036: _low_level_execute_command() done: rc=0, stdout=, stderr= 10562 1726776672.40041: _low_level_execute_command(): starting 10562 1726776672.40051: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_stat.py && sleep 0' 10562 1726776672.56540: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 312, "inode": 159383749, "dev": 51713, "nlink": 1, "atime": 1726776658.092236, "mtime": 1726776657.1702242, "ctime": 1726776657.4132273, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": true, "xgrp": false, "woth": false, "roth": true, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "mimetype": "text/plain", "charset": "us-ascii", "version": "646792945", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10562 1726776672.57522: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10562 1726776672.57570: stderr chunk (state=3): >>><<< 10562 1726776672.57578: stdout chunk (state=3): >>><<< 10562 1726776672.57594: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 312, "inode": 159383749, "dev": 51713, "nlink": 1, "atime": 1726776658.092236, "mtime": 1726776657.1702242, "ctime": 1726776657.4132273, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": true, "xgrp": false, "woth": false, "roth": true, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "mimetype": "text/plain", "charset": "us-ascii", "version": "646792945", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10562 1726776672.57634: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/kernel_settings/tuned.conf', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10562 1726776672.57666: variable 'ansible_module_compression' from source: unknown 10562 1726776672.57698: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10562 1726776672.57717: variable 'ansible_facts' from source: unknown 10562 1726776672.57775: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_file.py 10562 1726776672.57867: Sending initial data 10562 1726776672.57874: Sent initial data (151 bytes) 10562 1726776672.60404: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmps356tz9g /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_file.py <<< 10562 1726776672.61719: stderr chunk (state=3): >>><<< 10562 1726776672.61727: stdout chunk (state=3): >>><<< 10562 1726776672.61749: done transferring module to remote 10562 1726776672.61759: _low_level_execute_command(): starting 10562 1726776672.61765: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/ /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_file.py && sleep 0' 10562 1726776672.65035: stderr chunk (state=2): >>><<< 10562 1726776672.65048: stdout chunk (state=2): >>><<< 10562 1726776672.65065: _low_level_execute_command() done: rc=0, stdout=, stderr= 10562 1726776672.65070: _low_level_execute_command(): starting 10562 1726776672.65075: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/AnsiballZ_file.py && sleep 0' 10562 1726776672.81576: stdout chunk (state=2): >>> {"path": "/etc/tuned/kernel_settings/tuned.conf", "changed": false, "diff": {"before": {"path": "/etc/tuned/kernel_settings/tuned.conf"}, "after": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0644", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 312, "invocation": {"module_args": {"mode": "0644", "dest": "/etc/tuned/kernel_settings/tuned.conf", "_original_basename": "kernel_settings.j2", "recurse": false, "state": "file", "path": "/etc/tuned/kernel_settings/tuned.conf", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10562 1726776672.82721: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10562 1726776672.82772: stderr chunk (state=3): >>><<< 10562 1726776672.82779: stdout chunk (state=3): >>><<< 10562 1726776672.82795: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/kernel_settings/tuned.conf", "changed": false, "diff": {"before": {"path": "/etc/tuned/kernel_settings/tuned.conf"}, "after": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0644", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 312, "invocation": {"module_args": {"mode": "0644", "dest": "/etc/tuned/kernel_settings/tuned.conf", "_original_basename": "kernel_settings.j2", "recurse": false, "state": "file", "path": "/etc/tuned/kernel_settings/tuned.conf", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10562 1726776672.82823: done with _execute_module (ansible.legacy.file, {'mode': '0644', 'dest': '/etc/tuned/kernel_settings/tuned.conf', '_original_basename': 'kernel_settings.j2', 'recurse': False, 'state': 'file', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10562 1726776672.82852: _low_level_execute_command(): starting 10562 1726776672.82861: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776672.2426186-10562-26494737368727/ > /dev/null 2>&1 && sleep 0' 10562 1726776672.85262: stderr chunk (state=2): >>><<< 10562 1726776672.85271: stdout chunk (state=2): >>><<< 10562 1726776672.85285: _low_level_execute_command() done: rc=0, stdout=, stderr= 10562 1726776672.85296: handler run complete 10562 1726776672.85316: attempt loop complete, returning result 10562 1726776672.85319: _execute() done 10562 1726776672.85323: dumping result to json 10562 1726776672.85330: done dumping result, returning 10562 1726776672.85337: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings [120fa90a-8a95-fee7-0cea-00000000040f] 10562 1726776672.85343: sending task result for task 120fa90a-8a95-fee7-0cea-00000000040f 10562 1726776672.85388: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000040f 10562 1726776672.85392: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "dest": "/etc/tuned/kernel_settings/tuned.conf", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "path": "/etc/tuned/kernel_settings/tuned.conf", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 312, "state": "file", "uid": 0 } 9704 1726776672.85579: no more pending results, returning what we have 9704 1726776672.85582: results queue empty 9704 1726776672.85583: checking for any_errors_fatal 9704 1726776672.85590: done checking for any_errors_fatal 9704 1726776672.85590: checking for max_fail_percentage 9704 1726776672.85592: done checking for max_fail_percentage 9704 1726776672.85592: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.85593: done checking to see if all hosts have failed 9704 1726776672.85593: getting the remaining hosts for this loop 9704 1726776672.85595: done getting the remaining hosts for this loop 9704 1726776672.85598: getting the next task for host managed_node3 9704 1726776672.85603: done getting next task for host managed_node3 9704 1726776672.85607: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes 9704 1726776672.85609: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.85618: getting variables 9704 1726776672.85619: in VariableManager get_vars() 9704 1726776672.85655: Calling all_inventory to load vars for managed_node3 9704 1726776672.85658: Calling groups_inventory to load vars for managed_node3 9704 1726776672.85660: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.85668: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.85670: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.85673: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.85786: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.85911: done with get_vars() 9704 1726776672.85920: done getting variables 9704 1726776672.85969: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:149 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.657) 0:00:25.788 **** 9704 1726776672.85991: entering _queue_task() for managed_node3/service 9704 1726776672.86161: worker is 1 (out of 1 available) 9704 1726776672.86177: exiting _queue_task() for managed_node3/service 9704 1726776672.86189: done queuing things up, now waiting for results queue to drain 9704 1726776672.86191: waiting for pending results... 10593 1726776672.86310: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes 10593 1726776672.86421: in run() - task 120fa90a-8a95-fee7-0cea-000000000410 10593 1726776672.86438: variable 'ansible_search_path' from source: unknown 10593 1726776672.86442: variable 'ansible_search_path' from source: unknown 10593 1726776672.86477: variable '__kernel_settings_services' from source: include_vars 10593 1726776672.86713: variable '__kernel_settings_services' from source: include_vars 10593 1726776672.86843: variable 'omit' from source: magic vars 10593 1726776672.86908: variable 'ansible_host' from source: host vars for 'managed_node3' 10593 1726776672.86916: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10593 1726776672.86922: variable 'omit' from source: magic vars 10593 1726776672.87093: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10593 1726776672.87263: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10593 1726776672.87296: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10593 1726776672.87322: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10593 1726776672.87351: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10593 1726776672.87422: variable '__kernel_settings_register_profile' from source: set_fact 10593 1726776672.87436: variable '__kernel_settings_register_mode' from source: set_fact 10593 1726776672.87454: Evaluated conditional (__kernel_settings_register_profile is changed or __kernel_settings_register_mode is changed): False 10593 1726776672.87459: when evaluation is False, skipping this task 10593 1726776672.87478: variable 'item' from source: unknown 10593 1726776672.87524: variable 'item' from source: unknown skipping: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__kernel_settings_register_profile is changed or __kernel_settings_register_mode is changed", "item": "tuned", "skip_reason": "Conditional result was False" } 10593 1726776672.87553: dumping result to json 10593 1726776672.87559: done dumping result, returning 10593 1726776672.87564: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes [120fa90a-8a95-fee7-0cea-000000000410] 10593 1726776672.87570: sending task result for task 120fa90a-8a95-fee7-0cea-000000000410 10593 1726776672.87592: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000410 10593 1726776672.87595: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false } MSG: All items skipped 9704 1726776672.87754: no more pending results, returning what we have 9704 1726776672.87757: results queue empty 9704 1726776672.87757: checking for any_errors_fatal 9704 1726776672.87767: done checking for any_errors_fatal 9704 1726776672.87768: checking for max_fail_percentage 9704 1726776672.87769: done checking for max_fail_percentage 9704 1726776672.87770: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.87771: done checking to see if all hosts have failed 9704 1726776672.87771: getting the remaining hosts for this loop 9704 1726776672.87772: done getting the remaining hosts for this loop 9704 1726776672.87775: getting the next task for host managed_node3 9704 1726776672.87779: done getting next task for host managed_node3 9704 1726776672.87782: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings 9704 1726776672.87783: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.87793: getting variables 9704 1726776672.87794: in VariableManager get_vars() 9704 1726776672.87816: Calling all_inventory to load vars for managed_node3 9704 1726776672.87818: Calling groups_inventory to load vars for managed_node3 9704 1726776672.87819: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.87825: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.87827: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.87830: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.87933: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.88057: done with get_vars() 9704 1726776672.88064: done getting variables 9704 1726776672.88102: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Tuned apply settings] ******** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:157 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.021) 0:00:25.810 **** 9704 1726776672.88124: entering _queue_task() for managed_node3/command 9704 1726776672.88283: worker is 1 (out of 1 available) 9704 1726776672.88296: exiting _queue_task() for managed_node3/command 9704 1726776672.88306: done queuing things up, now waiting for results queue to drain 9704 1726776672.88308: waiting for pending results... 10594 1726776672.88427: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings 10594 1726776672.88537: in run() - task 120fa90a-8a95-fee7-0cea-000000000411 10594 1726776672.88553: variable 'ansible_search_path' from source: unknown 10594 1726776672.88558: variable 'ansible_search_path' from source: unknown 10594 1726776672.88584: calling self._execute() 10594 1726776672.88647: variable 'ansible_host' from source: host vars for 'managed_node3' 10594 1726776672.88655: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10594 1726776672.88663: variable 'omit' from source: magic vars 10594 1726776672.88980: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10594 1726776672.89247: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10594 1726776672.89279: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10594 1726776672.89312: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10594 1726776672.89340: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10594 1726776672.89421: variable '__kernel_settings_register_profile' from source: set_fact 10594 1726776672.89444: Evaluated conditional (not __kernel_settings_register_profile is changed): True 10594 1726776672.89535: variable '__kernel_settings_register_mode' from source: set_fact 10594 1726776672.89547: Evaluated conditional (not __kernel_settings_register_mode is changed): True 10594 1726776672.89619: variable '__kernel_settings_register_apply' from source: set_fact 10594 1726776672.89632: Evaluated conditional (__kernel_settings_register_apply is changed): False 10594 1726776672.89636: when evaluation is False, skipping this task 10594 1726776672.89639: _execute() done 10594 1726776672.89643: dumping result to json 10594 1726776672.89647: done dumping result, returning 10594 1726776672.89652: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings [120fa90a-8a95-fee7-0cea-000000000411] 10594 1726776672.89657: sending task result for task 120fa90a-8a95-fee7-0cea-000000000411 10594 1726776672.89679: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000411 10594 1726776672.89683: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_apply is changed", "skip_reason": "Conditional result was False" } 9704 1726776672.89785: no more pending results, returning what we have 9704 1726776672.89788: results queue empty 9704 1726776672.89789: checking for any_errors_fatal 9704 1726776672.89795: done checking for any_errors_fatal 9704 1726776672.89796: checking for max_fail_percentage 9704 1726776672.89798: done checking for max_fail_percentage 9704 1726776672.89798: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.89799: done checking to see if all hosts have failed 9704 1726776672.89799: getting the remaining hosts for this loop 9704 1726776672.89802: done getting the remaining hosts for this loop 9704 1726776672.89805: getting the next task for host managed_node3 9704 1726776672.89811: done getting next task for host managed_node3 9704 1726776672.89815: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Verify settings 9704 1726776672.89817: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.89832: getting variables 9704 1726776672.89834: in VariableManager get_vars() 9704 1726776672.89867: Calling all_inventory to load vars for managed_node3 9704 1726776672.89870: Calling groups_inventory to load vars for managed_node3 9704 1726776672.89871: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.89879: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.89882: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.89884: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.89992: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.90177: done with get_vars() 9704 1726776672.90184: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Verify settings] ************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:166 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.021) 0:00:25.831 **** 9704 1726776672.90251: entering _queue_task() for managed_node3/include_tasks 9704 1726776672.90410: worker is 1 (out of 1 available) 9704 1726776672.90424: exiting _queue_task() for managed_node3/include_tasks 9704 1726776672.90436: done queuing things up, now waiting for results queue to drain 9704 1726776672.90438: waiting for pending results... 10595 1726776672.90557: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Verify settings 10595 1726776672.90673: in run() - task 120fa90a-8a95-fee7-0cea-000000000412 10595 1726776672.90688: variable 'ansible_search_path' from source: unknown 10595 1726776672.90692: variable 'ansible_search_path' from source: unknown 10595 1726776672.90718: calling self._execute() 10595 1726776672.90784: variable 'ansible_host' from source: host vars for 'managed_node3' 10595 1726776672.90792: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10595 1726776672.90800: variable 'omit' from source: magic vars 10595 1726776672.91119: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10595 1726776672.91297: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10595 1726776672.91332: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10595 1726776672.91359: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10595 1726776672.91386: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10595 1726776672.91468: variable '__kernel_settings_register_apply' from source: set_fact 10595 1726776672.91491: Evaluated conditional (__kernel_settings_register_apply is changed): False 10595 1726776672.91496: when evaluation is False, skipping this task 10595 1726776672.91499: _execute() done 10595 1726776672.91503: dumping result to json 10595 1726776672.91506: done dumping result, returning 10595 1726776672.91512: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Verify settings [120fa90a-8a95-fee7-0cea-000000000412] 10595 1726776672.91517: sending task result for task 120fa90a-8a95-fee7-0cea-000000000412 10595 1726776672.91543: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000412 10595 1726776672.91549: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_apply is changed", "skip_reason": "Conditional result was False" } 9704 1726776672.91666: no more pending results, returning what we have 9704 1726776672.91669: results queue empty 9704 1726776672.91670: checking for any_errors_fatal 9704 1726776672.91675: done checking for any_errors_fatal 9704 1726776672.91676: checking for max_fail_percentage 9704 1726776672.91677: done checking for max_fail_percentage 9704 1726776672.91678: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.91678: done checking to see if all hosts have failed 9704 1726776672.91679: getting the remaining hosts for this loop 9704 1726776672.91680: done getting the remaining hosts for this loop 9704 1726776672.91683: getting the next task for host managed_node3 9704 1726776672.91689: done getting next task for host managed_node3 9704 1726776672.91692: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes 9704 1726776672.91694: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.91708: getting variables 9704 1726776672.91710: in VariableManager get_vars() 9704 1726776672.91741: Calling all_inventory to load vars for managed_node3 9704 1726776672.91743: Calling groups_inventory to load vars for managed_node3 9704 1726776672.91746: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.91753: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.91754: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.91756: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.91863: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.91984: done with get_vars() 9704 1726776672.91991: done getting variables 9704 1726776672.92031: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:177 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.018) 0:00:25.849 **** 9704 1726776672.92054: entering _queue_task() for managed_node3/set_fact 9704 1726776672.92208: worker is 1 (out of 1 available) 9704 1726776672.92221: exiting _queue_task() for managed_node3/set_fact 9704 1726776672.92235: done queuing things up, now waiting for results queue to drain 9704 1726776672.92236: waiting for pending results... 10596 1726776672.92356: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes 10596 1726776672.92469: in run() - task 120fa90a-8a95-fee7-0cea-000000000413 10596 1726776672.92484: variable 'ansible_search_path' from source: unknown 10596 1726776672.92488: variable 'ansible_search_path' from source: unknown 10596 1726776672.92514: calling self._execute() 10596 1726776672.92580: variable 'ansible_host' from source: host vars for 'managed_node3' 10596 1726776672.92588: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10596 1726776672.92596: variable 'omit' from source: magic vars 10596 1726776672.92682: variable 'omit' from source: magic vars 10596 1726776672.92715: variable 'omit' from source: magic vars 10596 1726776672.92742: variable 'omit' from source: magic vars 10596 1726776672.92774: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10596 1726776672.92800: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10596 1726776672.92819: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10596 1726776672.92836: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10596 1726776672.92848: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10596 1726776672.92872: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10596 1726776672.92876: variable 'ansible_host' from source: host vars for 'managed_node3' 10596 1726776672.92880: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10596 1726776672.92949: Set connection var ansible_module_compression to ZIP_DEFLATED 10596 1726776672.92957: Set connection var ansible_shell_type to sh 10596 1726776672.92965: Set connection var ansible_timeout to 10 10596 1726776672.92970: Set connection var ansible_pipelining to False 10596 1726776672.92977: Set connection var ansible_shell_executable to /bin/sh 10596 1726776672.92980: Set connection var ansible_connection to ssh 10596 1726776672.92996: variable 'ansible_shell_executable' from source: unknown 10596 1726776672.93000: variable 'ansible_connection' from source: unknown 10596 1726776672.93005: variable 'ansible_module_compression' from source: unknown 10596 1726776672.93008: variable 'ansible_shell_type' from source: unknown 10596 1726776672.93012: variable 'ansible_shell_executable' from source: unknown 10596 1726776672.93015: variable 'ansible_host' from source: host vars for 'managed_node3' 10596 1726776672.93019: variable 'ansible_pipelining' from source: unknown 10596 1726776672.93023: variable 'ansible_timeout' from source: unknown 10596 1726776672.93026: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10596 1726776672.93112: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10596 1726776672.93121: variable 'omit' from source: magic vars 10596 1726776672.93126: starting attempt loop 10596 1726776672.93128: running the handler 10596 1726776672.93138: handler run complete 10596 1726776672.93147: attempt loop complete, returning result 10596 1726776672.93151: _execute() done 10596 1726776672.93154: dumping result to json 10596 1726776672.93157: done dumping result, returning 10596 1726776672.93163: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes [120fa90a-8a95-fee7-0cea-000000000413] 10596 1726776672.93169: sending task result for task 120fa90a-8a95-fee7-0cea-000000000413 10596 1726776672.93186: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000413 10596 1726776672.93189: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "kernel_settings_reboot_required": false }, "changed": false } 9704 1726776672.93335: no more pending results, returning what we have 9704 1726776672.93338: results queue empty 9704 1726776672.93338: checking for any_errors_fatal 9704 1726776672.93344: done checking for any_errors_fatal 9704 1726776672.93347: checking for max_fail_percentage 9704 1726776672.93349: done checking for max_fail_percentage 9704 1726776672.93350: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.93350: done checking to see if all hosts have failed 9704 1726776672.93351: getting the remaining hosts for this loop 9704 1726776672.93352: done getting the remaining hosts for this loop 9704 1726776672.93355: getting the next task for host managed_node3 9704 1726776672.93360: done getting next task for host managed_node3 9704 1726776672.93363: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing 9704 1726776672.93366: ^ state is: HOST STATE: block=2, task=18, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.93376: getting variables 9704 1726776672.93377: in VariableManager get_vars() 9704 1726776672.93406: Calling all_inventory to load vars for managed_node3 9704 1726776672.93409: Calling groups_inventory to load vars for managed_node3 9704 1726776672.93411: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.93417: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.93419: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.93420: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.93572: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.93689: done with get_vars() 9704 1726776672.93695: done getting variables 9704 1726776672.93755: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:181 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.017) 0:00:25.866 **** 9704 1726776672.93782: entering _queue_task() for managed_node3/set_fact 9704 1726776672.93981: worker is 1 (out of 1 available) 9704 1726776672.93996: exiting _queue_task() for managed_node3/set_fact 9704 1726776672.94008: done queuing things up, now waiting for results queue to drain 9704 1726776672.94010: waiting for pending results... 10597 1726776672.94237: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing 10597 1726776672.94374: in run() - task 120fa90a-8a95-fee7-0cea-000000000414 10597 1726776672.94393: variable 'ansible_search_path' from source: unknown 10597 1726776672.94397: variable 'ansible_search_path' from source: unknown 10597 1726776672.94431: calling self._execute() 10597 1726776672.94520: variable 'ansible_host' from source: host vars for 'managed_node3' 10597 1726776672.94531: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10597 1726776672.94541: variable 'omit' from source: magic vars 10597 1726776672.94649: variable 'omit' from source: magic vars 10597 1726776672.94686: variable 'omit' from source: magic vars 10597 1726776672.94960: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10597 1726776672.95149: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10597 1726776672.95184: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10597 1726776672.95212: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10597 1726776672.95242: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10597 1726776672.95351: variable '__kernel_settings_register_profile' from source: set_fact 10597 1726776672.95364: variable '__kernel_settings_register_mode' from source: set_fact 10597 1726776672.95372: variable '__kernel_settings_register_apply' from source: set_fact 10597 1726776672.95410: variable 'omit' from source: magic vars 10597 1726776672.95435: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10597 1726776672.95459: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10597 1726776672.95475: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10597 1726776672.95489: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10597 1726776672.95497: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10597 1726776672.95520: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10597 1726776672.95524: variable 'ansible_host' from source: host vars for 'managed_node3' 10597 1726776672.95526: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10597 1726776672.95598: Set connection var ansible_module_compression to ZIP_DEFLATED 10597 1726776672.95606: Set connection var ansible_shell_type to sh 10597 1726776672.95613: Set connection var ansible_timeout to 10 10597 1726776672.95617: Set connection var ansible_pipelining to False 10597 1726776672.95621: Set connection var ansible_shell_executable to /bin/sh 10597 1726776672.95623: Set connection var ansible_connection to ssh 10597 1726776672.95638: variable 'ansible_shell_executable' from source: unknown 10597 1726776672.95642: variable 'ansible_connection' from source: unknown 10597 1726776672.95644: variable 'ansible_module_compression' from source: unknown 10597 1726776672.95648: variable 'ansible_shell_type' from source: unknown 10597 1726776672.95650: variable 'ansible_shell_executable' from source: unknown 10597 1726776672.95652: variable 'ansible_host' from source: host vars for 'managed_node3' 10597 1726776672.95654: variable 'ansible_pipelining' from source: unknown 10597 1726776672.95656: variable 'ansible_timeout' from source: unknown 10597 1726776672.95658: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10597 1726776672.95722: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10597 1726776672.95733: variable 'omit' from source: magic vars 10597 1726776672.95738: starting attempt loop 10597 1726776672.95742: running the handler 10597 1726776672.95753: handler run complete 10597 1726776672.95759: attempt loop complete, returning result 10597 1726776672.95761: _execute() done 10597 1726776672.95763: dumping result to json 10597 1726776672.95765: done dumping result, returning 10597 1726776672.95769: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing [120fa90a-8a95-fee7-0cea-000000000414] 10597 1726776672.95773: sending task result for task 120fa90a-8a95-fee7-0cea-000000000414 10597 1726776672.95788: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000414 10597 1726776672.95790: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_changed": false }, "changed": false } 9704 1726776672.96088: no more pending results, returning what we have 9704 1726776672.96090: results queue empty 9704 1726776672.96091: checking for any_errors_fatal 9704 1726776672.96095: done checking for any_errors_fatal 9704 1726776672.96095: checking for max_fail_percentage 9704 1726776672.96097: done checking for max_fail_percentage 9704 1726776672.96097: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.96098: done checking to see if all hosts have failed 9704 1726776672.96098: getting the remaining hosts for this loop 9704 1726776672.96099: done getting the remaining hosts for this loop 9704 1726776672.96101: getting the next task for host managed_node3 9704 1726776672.96107: done getting next task for host managed_node3 9704 1726776672.96108: ^ task is: TASK: meta (role_complete) 9704 1726776672.96110: ^ state is: HOST STATE: block=2, task=19, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.96117: getting variables 9704 1726776672.96118: in VariableManager get_vars() 9704 1726776672.96148: Calling all_inventory to load vars for managed_node3 9704 1726776672.96151: Calling groups_inventory to load vars for managed_node3 9704 1726776672.96153: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.96159: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.96161: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.96162: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.96274: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.96397: done with get_vars() 9704 1726776672.96406: done getting variables 9704 1726776672.96464: done queuing things up, now waiting for results queue to drain 9704 1726776672.96465: results queue empty 9704 1726776672.96465: checking for any_errors_fatal 9704 1726776672.96468: done checking for any_errors_fatal 9704 1726776672.96468: checking for max_fail_percentage 9704 1726776672.96469: done checking for max_fail_percentage 9704 1726776672.96474: checking to see if all hosts have failed and the running result is not ok 9704 1726776672.96474: done checking to see if all hosts have failed 9704 1726776672.96475: getting the remaining hosts for this loop 9704 1726776672.96475: done getting the remaining hosts for this loop 9704 1726776672.96477: getting the next task for host managed_node3 9704 1726776672.96479: done getting next task for host managed_node3 9704 1726776672.96480: ^ task is: TASK: Ensure role reported not changed 9704 1726776672.96481: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776672.96483: getting variables 9704 1726776672.96483: in VariableManager get_vars() 9704 1726776672.96491: Calling all_inventory to load vars for managed_node3 9704 1726776672.96492: Calling groups_inventory to load vars for managed_node3 9704 1726776672.96493: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776672.96496: Calling all_plugins_play to load vars for managed_node3 9704 1726776672.96497: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776672.96499: Calling groups_plugins_play to load vars for managed_node3 9704 1726776672.96624: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776672.96828: done with get_vars() 9704 1726776672.96837: done getting variables 9704 1726776672.96872: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Ensure role reported not changed] **************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:67 Thursday 19 September 2024 16:11:12 -0400 (0:00:00.031) 0:00:25.897 **** 9704 1726776672.96896: entering _queue_task() for managed_node3/assert 9704 1726776672.97115: worker is 1 (out of 1 available) 9704 1726776672.97132: exiting _queue_task() for managed_node3/assert 9704 1726776672.97147: done queuing things up, now waiting for results queue to drain 9704 1726776672.97150: waiting for pending results... 10599 1726776672.97366: running TaskExecutor() for managed_node3/TASK: Ensure role reported not changed 10599 1726776672.97475: in run() - task 120fa90a-8a95-fee7-0cea-00000000000f 10599 1726776672.97495: variable 'ansible_search_path' from source: unknown 10599 1726776672.97527: calling self._execute() 10599 1726776672.97619: variable 'ansible_host' from source: host vars for 'managed_node3' 10599 1726776672.97631: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10599 1726776672.97640: variable 'omit' from source: magic vars 10599 1726776672.97744: variable 'omit' from source: magic vars 10599 1726776672.97777: variable 'omit' from source: magic vars 10599 1726776672.97809: variable 'omit' from source: magic vars 10599 1726776672.97852: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10599 1726776672.97884: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10599 1726776672.97905: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10599 1726776672.97924: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10599 1726776672.97941: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10599 1726776672.97975: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10599 1726776672.97981: variable 'ansible_host' from source: host vars for 'managed_node3' 10599 1726776672.97984: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10599 1726776672.98087: Set connection var ansible_module_compression to ZIP_DEFLATED 10599 1726776672.98096: Set connection var ansible_shell_type to sh 10599 1726776672.98104: Set connection var ansible_timeout to 10 10599 1726776672.98110: Set connection var ansible_pipelining to False 10599 1726776672.98117: Set connection var ansible_shell_executable to /bin/sh 10599 1726776672.98120: Set connection var ansible_connection to ssh 10599 1726776672.98144: variable 'ansible_shell_executable' from source: unknown 10599 1726776672.98152: variable 'ansible_connection' from source: unknown 10599 1726776672.98155: variable 'ansible_module_compression' from source: unknown 10599 1726776672.98158: variable 'ansible_shell_type' from source: unknown 10599 1726776672.98161: variable 'ansible_shell_executable' from source: unknown 10599 1726776672.98164: variable 'ansible_host' from source: host vars for 'managed_node3' 10599 1726776672.98167: variable 'ansible_pipelining' from source: unknown 10599 1726776672.98170: variable 'ansible_timeout' from source: unknown 10599 1726776672.98174: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10599 1726776672.98397: Loading ActionModule 'assert' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/assert.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10599 1726776672.98411: variable 'omit' from source: magic vars 10599 1726776672.98417: starting attempt loop 10599 1726776672.98421: running the handler 10599 1726776672.98791: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10599 1726776673.01154: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10599 1726776673.01225: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10599 1726776673.01266: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10599 1726776673.01301: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10599 1726776673.01327: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10599 1726776673.01398: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10599 1726776673.01428: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10599 1726776673.01459: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10599 1726776673.01499: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10599 1726776673.01513: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10599 1726776673.01663: variable '__kernel_settings_changed' from source: set_fact 10599 1726776673.01682: Evaluated conditional (not __kernel_settings_changed | d(false)): True 10599 1726776673.01690: handler run complete 10599 1726776673.01714: attempt loop complete, returning result 10599 1726776673.01718: _execute() done 10599 1726776673.01722: dumping result to json 10599 1726776673.01726: done dumping result, returning 10599 1726776673.01734: done running TaskExecutor() for managed_node3/TASK: Ensure role reported not changed [120fa90a-8a95-fee7-0cea-00000000000f] 10599 1726776673.01741: sending task result for task 120fa90a-8a95-fee7-0cea-00000000000f 10599 1726776673.01773: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000000f 10599 1726776673.01777: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false } MSG: All assertions passed 9704 1726776673.02202: no more pending results, returning what we have 9704 1726776673.02206: results queue empty 9704 1726776673.02206: checking for any_errors_fatal 9704 1726776673.02208: done checking for any_errors_fatal 9704 1726776673.02209: checking for max_fail_percentage 9704 1726776673.02211: done checking for max_fail_percentage 9704 1726776673.02211: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.02212: done checking to see if all hosts have failed 9704 1726776673.02213: getting the remaining hosts for this loop 9704 1726776673.02214: done getting the remaining hosts for this loop 9704 1726776673.02217: getting the next task for host managed_node3 9704 1726776673.02224: done getting next task for host managed_node3 9704 1726776673.02226: ^ task is: TASK: Cleanup 9704 1726776673.02228: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776673.02233: getting variables 9704 1726776673.02235: in VariableManager get_vars() 9704 1726776673.02271: Calling all_inventory to load vars for managed_node3 9704 1726776673.02274: Calling groups_inventory to load vars for managed_node3 9704 1726776673.02276: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.02286: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.02289: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.02298: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.02477: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.02680: done with get_vars() 9704 1726776673.02690: done getting variables TASK [Cleanup] ***************************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:72 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.058) 0:00:25.956 **** 9704 1726776673.02776: entering _queue_task() for managed_node3/include_tasks 9704 1726776673.02953: worker is 1 (out of 1 available) 9704 1726776673.02968: exiting _queue_task() for managed_node3/include_tasks 9704 1726776673.02980: done queuing things up, now waiting for results queue to drain 9704 1726776673.02981: waiting for pending results... 10601 1726776673.03119: running TaskExecutor() for managed_node3/TASK: Cleanup 10601 1726776673.03236: in run() - task 120fa90a-8a95-fee7-0cea-000000000010 10601 1726776673.03254: variable 'ansible_search_path' from source: unknown 10601 1726776673.03288: calling self._execute() 10601 1726776673.03376: variable 'ansible_host' from source: host vars for 'managed_node3' 10601 1726776673.03385: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10601 1726776673.03394: variable 'omit' from source: magic vars 10601 1726776673.03491: _execute() done 10601 1726776673.03497: dumping result to json 10601 1726776673.03501: done dumping result, returning 10601 1726776673.03506: done running TaskExecutor() for managed_node3/TASK: Cleanup [120fa90a-8a95-fee7-0cea-000000000010] 10601 1726776673.03512: sending task result for task 120fa90a-8a95-fee7-0cea-000000000010 10601 1726776673.03541: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000010 10601 1726776673.03545: WORKER PROCESS EXITING 9704 1726776673.03852: no more pending results, returning what we have 9704 1726776673.03856: in VariableManager get_vars() 9704 1726776673.03891: Calling all_inventory to load vars for managed_node3 9704 1726776673.03894: Calling groups_inventory to load vars for managed_node3 9704 1726776673.03896: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.03905: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.03908: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.03911: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.04074: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.04296: done with get_vars() 9704 1726776673.04303: variable 'ansible_search_path' from source: unknown 9704 1726776673.04317: we have included files to process 9704 1726776673.04318: generating all_blocks data 9704 1726776673.04323: done generating all_blocks data 9704 1726776673.04326: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml 9704 1726776673.04327: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml 9704 1726776673.04331: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml for managed_node3 9704 1726776673.05262: done processing included file 9704 1726776673.05264: iterating over new_blocks loaded from include file 9704 1726776673.05266: in VariableManager get_vars() 9704 1726776673.05281: done with get_vars() 9704 1726776673.05282: filtering new block on tags 9704 1726776673.05305: done filtering new block on tags 9704 1726776673.05307: in VariableManager get_vars() 9704 1726776673.05320: done with get_vars() 9704 1726776673.05322: filtering new block on tags 9704 1726776673.05412: done filtering new block on tags 9704 1726776673.05415: done iterating over new_blocks loaded from include file 9704 1726776673.05416: extending task lists for all hosts with included blocks 9704 1726776673.07438: done extending task lists 9704 1726776673.07439: done processing included files 9704 1726776673.07439: results queue empty 9704 1726776673.07440: checking for any_errors_fatal 9704 1726776673.07442: done checking for any_errors_fatal 9704 1726776673.07442: checking for max_fail_percentage 9704 1726776673.07443: done checking for max_fail_percentage 9704 1726776673.07443: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.07444: done checking to see if all hosts have failed 9704 1726776673.07444: getting the remaining hosts for this loop 9704 1726776673.07445: done getting the remaining hosts for this loop 9704 1726776673.07447: getting the next task for host managed_node3 9704 1726776673.07450: done getting next task for host managed_node3 9704 1726776673.07451: ^ task is: TASK: Show current tuned profile settings 9704 1726776673.07453: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.07454: getting variables 9704 1726776673.07455: in VariableManager get_vars() 9704 1726776673.07462: Calling all_inventory to load vars for managed_node3 9704 1726776673.07464: Calling groups_inventory to load vars for managed_node3 9704 1726776673.07465: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.07469: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.07470: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.07472: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.07567: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.07672: done with get_vars() 9704 1726776673.07678: done getting variables 9704 1726776673.07702: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Show current tuned profile settings] ************************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:2 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.049) 0:00:26.006 **** 9704 1726776673.07721: entering _queue_task() for managed_node3/command 9704 1726776673.07901: worker is 1 (out of 1 available) 9704 1726776673.07915: exiting _queue_task() for managed_node3/command 9704 1726776673.07928: done queuing things up, now waiting for results queue to drain 9704 1726776673.07931: waiting for pending results... 10604 1726776673.08052: running TaskExecutor() for managed_node3/TASK: Show current tuned profile settings 10604 1726776673.08156: in run() - task 120fa90a-8a95-fee7-0cea-000000000588 10604 1726776673.08173: variable 'ansible_search_path' from source: unknown 10604 1726776673.08177: variable 'ansible_search_path' from source: unknown 10604 1726776673.08205: calling self._execute() 10604 1726776673.08271: variable 'ansible_host' from source: host vars for 'managed_node3' 10604 1726776673.08280: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10604 1726776673.08288: variable 'omit' from source: magic vars 10604 1726776673.08363: variable 'omit' from source: magic vars 10604 1726776673.08392: variable 'omit' from source: magic vars 10604 1726776673.08621: variable '__kernel_settings_profile_filename' from source: role '' exported vars 10604 1726776673.08681: variable '__kernel_settings_profile_dir' from source: role '' exported vars 10604 1726776673.08744: variable '__kernel_settings_profile_parent' from source: set_fact 10604 1726776673.08755: variable '__kernel_settings_tuned_profile' from source: role '' exported vars 10604 1726776673.08788: variable 'omit' from source: magic vars 10604 1726776673.08819: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10604 1726776673.08848: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10604 1726776673.08868: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10604 1726776673.08882: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10604 1726776673.08893: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10604 1726776673.08918: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10604 1726776673.08923: variable 'ansible_host' from source: host vars for 'managed_node3' 10604 1726776673.08927: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10604 1726776673.08993: Set connection var ansible_module_compression to ZIP_DEFLATED 10604 1726776673.09002: Set connection var ansible_shell_type to sh 10604 1726776673.09013: Set connection var ansible_timeout to 10 10604 1726776673.09023: Set connection var ansible_pipelining to False 10604 1726776673.09031: Set connection var ansible_shell_executable to /bin/sh 10604 1726776673.09035: Set connection var ansible_connection to ssh 10604 1726776673.09056: variable 'ansible_shell_executable' from source: unknown 10604 1726776673.09061: variable 'ansible_connection' from source: unknown 10604 1726776673.09065: variable 'ansible_module_compression' from source: unknown 10604 1726776673.09069: variable 'ansible_shell_type' from source: unknown 10604 1726776673.09072: variable 'ansible_shell_executable' from source: unknown 10604 1726776673.09076: variable 'ansible_host' from source: host vars for 'managed_node3' 10604 1726776673.09083: variable 'ansible_pipelining' from source: unknown 10604 1726776673.09087: variable 'ansible_timeout' from source: unknown 10604 1726776673.09089: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10604 1726776673.09186: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10604 1726776673.09195: variable 'omit' from source: magic vars 10604 1726776673.09199: starting attempt loop 10604 1726776673.09201: running the handler 10604 1726776673.09211: _low_level_execute_command(): starting 10604 1726776673.09218: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10604 1726776673.11631: stdout chunk (state=2): >>>/root <<< 10604 1726776673.11752: stderr chunk (state=3): >>><<< 10604 1726776673.11758: stdout chunk (state=3): >>><<< 10604 1726776673.11774: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10604 1726776673.11786: _low_level_execute_command(): starting 10604 1726776673.11791: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731 `" && echo ansible-tmp-1726776673.117812-10604-57674612362731="` echo /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731 `" ) && sleep 0' 10604 1726776673.14176: stdout chunk (state=2): >>>ansible-tmp-1726776673.117812-10604-57674612362731=/root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731 <<< 10604 1726776673.14297: stderr chunk (state=3): >>><<< 10604 1726776673.14303: stdout chunk (state=3): >>><<< 10604 1726776673.14318: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776673.117812-10604-57674612362731=/root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731 , stderr= 10604 1726776673.14344: variable 'ansible_module_compression' from source: unknown 10604 1726776673.14382: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 10604 1726776673.14407: variable 'ansible_facts' from source: unknown 10604 1726776673.14481: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/AnsiballZ_command.py 10604 1726776673.14574: Sending initial data 10604 1726776673.14581: Sent initial data (153 bytes) 10604 1726776673.17024: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpo2_niw5i /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/AnsiballZ_command.py <<< 10604 1726776673.17997: stderr chunk (state=3): >>><<< 10604 1726776673.18004: stdout chunk (state=3): >>><<< 10604 1726776673.18022: done transferring module to remote 10604 1726776673.18033: _low_level_execute_command(): starting 10604 1726776673.18039: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/ /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/AnsiballZ_command.py && sleep 0' 10604 1726776673.20362: stderr chunk (state=2): >>><<< 10604 1726776673.20370: stdout chunk (state=2): >>><<< 10604 1726776673.20384: _low_level_execute_command() done: rc=0, stdout=, stderr= 10604 1726776673.20389: _low_level_execute_command(): starting 10604 1726776673.20394: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/AnsiballZ_command.py && sleep 0' 10604 1726776673.35937: stdout chunk (state=2): >>> {"changed": true, "stdout": "#\n# Ansible managed\n#\n# system_role:kernel_settings\n\n[main]\nsummary = kernel settings\n[sysctl]\nfs.epoll.max_user_watches = 785592\nfs.file-max = 379724\n[sysfs]\n/sys/kernel/debug/x86/ibrs_enabled = 0\n/sys/kernel/debug/x86/pti_enabled = 0\n/sys/kernel/debug/x86/retp_enabled = 0\n[vm]\ntransparent_hugepages = madvise", "stderr": "", "rc": 0, "cmd": ["cat", "/etc/tuned/kernel_settings/tuned.conf"], "start": "2024-09-19 16:11:13.355002", "end": "2024-09-19 16:11:13.357777", "delta": "0:00:00.002775", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /etc/tuned/kernel_settings/tuned.conf", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 10604 1726776673.37191: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10604 1726776673.37202: stdout chunk (state=3): >>><<< 10604 1726776673.37214: stderr chunk (state=3): >>><<< 10604 1726776673.37230: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "#\n# Ansible managed\n#\n# system_role:kernel_settings\n\n[main]\nsummary = kernel settings\n[sysctl]\nfs.epoll.max_user_watches = 785592\nfs.file-max = 379724\n[sysfs]\n/sys/kernel/debug/x86/ibrs_enabled = 0\n/sys/kernel/debug/x86/pti_enabled = 0\n/sys/kernel/debug/x86/retp_enabled = 0\n[vm]\ntransparent_hugepages = madvise", "stderr": "", "rc": 0, "cmd": ["cat", "/etc/tuned/kernel_settings/tuned.conf"], "start": "2024-09-19 16:11:13.355002", "end": "2024-09-19 16:11:13.357777", "delta": "0:00:00.002775", "msg": "", "invocation": {"module_args": {"_raw_params": "cat /etc/tuned/kernel_settings/tuned.conf", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10604 1726776673.37275: done with _execute_module (ansible.legacy.command, {'_raw_params': 'cat /etc/tuned/kernel_settings/tuned.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10604 1726776673.37287: _low_level_execute_command(): starting 10604 1726776673.37292: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776673.117812-10604-57674612362731/ > /dev/null 2>&1 && sleep 0' 10604 1726776673.39849: stderr chunk (state=2): >>><<< 10604 1726776673.39857: stdout chunk (state=2): >>><<< 10604 1726776673.39875: _low_level_execute_command() done: rc=0, stdout=, stderr= 10604 1726776673.39882: handler run complete 10604 1726776673.39904: Evaluated conditional (False): False 10604 1726776673.39917: attempt loop complete, returning result 10604 1726776673.39925: _execute() done 10604 1726776673.39931: dumping result to json 10604 1726776673.39938: done dumping result, returning 10604 1726776673.39944: done running TaskExecutor() for managed_node3/TASK: Show current tuned profile settings [120fa90a-8a95-fee7-0cea-000000000588] 10604 1726776673.39953: sending task result for task 120fa90a-8a95-fee7-0cea-000000000588 10604 1726776673.39994: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000588 10604 1726776673.39998: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "cmd": [ "cat", "/etc/tuned/kernel_settings/tuned.conf" ], "delta": "0:00:00.002775", "end": "2024-09-19 16:11:13.357777", "rc": 0, "start": "2024-09-19 16:11:13.355002" } STDOUT: # # Ansible managed # # system_role:kernel_settings [main] summary = kernel settings [sysctl] fs.epoll.max_user_watches = 785592 fs.file-max = 379724 [sysfs] /sys/kernel/debug/x86/ibrs_enabled = 0 /sys/kernel/debug/x86/pti_enabled = 0 /sys/kernel/debug/x86/retp_enabled = 0 [vm] transparent_hugepages = madvise 9704 1726776673.40378: no more pending results, returning what we have 9704 1726776673.40381: results queue empty 9704 1726776673.40382: checking for any_errors_fatal 9704 1726776673.40384: done checking for any_errors_fatal 9704 1726776673.40386: checking for max_fail_percentage 9704 1726776673.40388: done checking for max_fail_percentage 9704 1726776673.40389: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.40390: done checking to see if all hosts have failed 9704 1726776673.40390: getting the remaining hosts for this loop 9704 1726776673.40391: done getting the remaining hosts for this loop 9704 1726776673.40395: getting the next task for host managed_node3 9704 1726776673.40405: done getting next task for host managed_node3 9704 1726776673.40407: ^ task is: TASK: Run role with purge to remove everything 9704 1726776673.40410: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.40413: getting variables 9704 1726776673.40415: in VariableManager get_vars() 9704 1726776673.40452: Calling all_inventory to load vars for managed_node3 9704 1726776673.40455: Calling groups_inventory to load vars for managed_node3 9704 1726776673.40456: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.40466: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.40468: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.40471: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.40642: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.40851: done with get_vars() 9704 1726776673.40864: done getting variables TASK [Run role with purge to remove everything] ******************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:9 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.332) 0:00:26.338 **** 9704 1726776673.40957: entering _queue_task() for managed_node3/include_role 9704 1726776673.41149: worker is 1 (out of 1 available) 9704 1726776673.41162: exiting _queue_task() for managed_node3/include_role 9704 1726776673.41173: done queuing things up, now waiting for results queue to drain 9704 1726776673.41175: waiting for pending results... 10619 1726776673.41296: running TaskExecutor() for managed_node3/TASK: Run role with purge to remove everything 10619 1726776673.41400: in run() - task 120fa90a-8a95-fee7-0cea-00000000058a 10619 1726776673.41416: variable 'ansible_search_path' from source: unknown 10619 1726776673.41420: variable 'ansible_search_path' from source: unknown 10619 1726776673.41451: calling self._execute() 10619 1726776673.41520: variable 'ansible_host' from source: host vars for 'managed_node3' 10619 1726776673.41529: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10619 1726776673.41539: variable 'omit' from source: magic vars 10619 1726776673.41613: _execute() done 10619 1726776673.41619: dumping result to json 10619 1726776673.41624: done dumping result, returning 10619 1726776673.41631: done running TaskExecutor() for managed_node3/TASK: Run role with purge to remove everything [120fa90a-8a95-fee7-0cea-00000000058a] 10619 1726776673.41639: sending task result for task 120fa90a-8a95-fee7-0cea-00000000058a 10619 1726776673.41668: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000058a 10619 1726776673.41672: WORKER PROCESS EXITING 9704 1726776673.41772: no more pending results, returning what we have 9704 1726776673.41776: in VariableManager get_vars() 9704 1726776673.41811: Calling all_inventory to load vars for managed_node3 9704 1726776673.41814: Calling groups_inventory to load vars for managed_node3 9704 1726776673.41816: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.41825: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.41827: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.41832: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.42019: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.42199: done with get_vars() 9704 1726776673.42207: variable 'ansible_search_path' from source: unknown 9704 1726776673.42208: variable 'ansible_search_path' from source: unknown 9704 1726776673.42510: variable 'omit' from source: magic vars 9704 1726776673.42543: variable 'omit' from source: magic vars 9704 1726776673.42557: variable 'omit' from source: magic vars 9704 1726776673.42560: we have included files to process 9704 1726776673.42561: generating all_blocks data 9704 1726776673.42562: done generating all_blocks data 9704 1726776673.42566: processing included file: fedora.linux_system_roles.kernel_settings 9704 1726776673.42588: in VariableManager get_vars() 9704 1726776673.42604: done with get_vars() 9704 1726776673.42633: in VariableManager get_vars() 9704 1726776673.42651: done with get_vars() 9704 1726776673.42689: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/vars/main.yml 9704 1726776673.42744: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/defaults/main.yml 9704 1726776673.42768: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/meta/main.yml 9704 1726776673.42815: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml 9704 1726776673.43146: in VariableManager get_vars() 9704 1726776673.43160: done with get_vars() 9704 1726776673.43982: in VariableManager get_vars() 9704 1726776673.43997: done with get_vars() 9704 1726776673.44100: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/handlers/main.yml 9704 1726776673.44497: iterating over new_blocks loaded from include file 9704 1726776673.44499: in VariableManager get_vars() 9704 1726776673.44511: done with get_vars() 9704 1726776673.44512: filtering new block on tags 9704 1726776673.44714: done filtering new block on tags 9704 1726776673.44716: in VariableManager get_vars() 9704 1726776673.44727: done with get_vars() 9704 1726776673.44728: filtering new block on tags 9704 1726776673.44769: done filtering new block on tags 9704 1726776673.44772: in VariableManager get_vars() 9704 1726776673.44785: done with get_vars() 9704 1726776673.44787: filtering new block on tags 9704 1726776673.44884: done filtering new block on tags 9704 1726776673.44886: in VariableManager get_vars() 9704 1726776673.44896: done with get_vars() 9704 1726776673.44897: filtering new block on tags 9704 1726776673.44909: done filtering new block on tags 9704 1726776673.44910: done iterating over new_blocks loaded from include file 9704 1726776673.44911: extending task lists for all hosts with included blocks 9704 1726776673.45239: done extending task lists 9704 1726776673.45241: done processing included files 9704 1726776673.45241: results queue empty 9704 1726776673.45242: checking for any_errors_fatal 9704 1726776673.45248: done checking for any_errors_fatal 9704 1726776673.45248: checking for max_fail_percentage 9704 1726776673.45249: done checking for max_fail_percentage 9704 1726776673.45250: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.45251: done checking to see if all hosts have failed 9704 1726776673.45251: getting the remaining hosts for this loop 9704 1726776673.45252: done getting the remaining hosts for this loop 9704 1726776673.45254: getting the next task for host managed_node3 9704 1726776673.45258: done getting next task for host managed_node3 9704 1726776673.45261: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values 9704 1726776673.45263: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.45271: getting variables 9704 1726776673.45272: in VariableManager get_vars() 9704 1726776673.45283: Calling all_inventory to load vars for managed_node3 9704 1726776673.45285: Calling groups_inventory to load vars for managed_node3 9704 1726776673.45287: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.45292: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.45294: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.45297: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.45432: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.45606: done with get_vars() 9704 1726776673.45616: done getting variables 9704 1726776673.45655: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:2 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.047) 0:00:26.385 **** 9704 1726776673.45681: entering _queue_task() for managed_node3/fail 9704 1726776673.45876: worker is 1 (out of 1 available) 9704 1726776673.45889: exiting _queue_task() for managed_node3/fail 9704 1726776673.45901: done queuing things up, now waiting for results queue to drain 9704 1726776673.45903: waiting for pending results... 10622 1726776673.46028: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values 10622 1726776673.46155: in run() - task 120fa90a-8a95-fee7-0cea-0000000006d8 10622 1726776673.46172: variable 'ansible_search_path' from source: unknown 10622 1726776673.46176: variable 'ansible_search_path' from source: unknown 10622 1726776673.46204: calling self._execute() 10622 1726776673.46272: variable 'ansible_host' from source: host vars for 'managed_node3' 10622 1726776673.46280: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10622 1726776673.46289: variable 'omit' from source: magic vars 10622 1726776673.46625: variable 'kernel_settings_sysctl' from source: include params 10622 1726776673.46637: variable '__kernel_settings_state_empty' from source: role '' all vars 10622 1726776673.46649: Evaluated conditional (kernel_settings_sysctl != __kernel_settings_state_empty): True 10622 1726776673.46845: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10622 1726776673.48372: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10622 1726776673.48421: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10622 1726776673.48454: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10622 1726776673.48490: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10622 1726776673.48513: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10622 1726776673.48570: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10622 1726776673.48592: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10622 1726776673.48613: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10622 1726776673.48643: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10622 1726776673.48658: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10622 1726776673.48695: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10622 1726776673.48712: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10622 1726776673.48733: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10622 1726776673.48762: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10622 1726776673.48773: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10622 1726776673.48801: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10622 1726776673.48818: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10622 1726776673.48837: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10622 1726776673.48863: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10622 1726776673.48871: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10622 1726776673.49061: variable 'kernel_settings_sysctl' from source: include params 10622 1726776673.49083: Evaluated conditional ((kernel_settings_sysctl | selectattr("value", "defined") | selectattr("value", "sameas", true) | list | length > 0) or (kernel_settings_sysctl | selectattr("value", "defined") | selectattr("value", "sameas", false) | list | length > 0)): False 10622 1726776673.49088: when evaluation is False, skipping this task 10622 1726776673.49091: _execute() done 10622 1726776673.49095: dumping result to json 10622 1726776673.49098: done dumping result, returning 10622 1726776673.49105: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check sysctl settings for boolean values [120fa90a-8a95-fee7-0cea-0000000006d8] 10622 1726776673.49110: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006d8 10622 1726776673.49135: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006d8 10622 1726776673.49139: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "(kernel_settings_sysctl | selectattr(\"value\", \"defined\") | selectattr(\"value\", \"sameas\", true) | list | length > 0) or (kernel_settings_sysctl | selectattr(\"value\", \"defined\") | selectattr(\"value\", \"sameas\", false) | list | length > 0)", "skip_reason": "Conditional result was False" } 9704 1726776673.49256: no more pending results, returning what we have 9704 1726776673.49259: results queue empty 9704 1726776673.49260: checking for any_errors_fatal 9704 1726776673.49262: done checking for any_errors_fatal 9704 1726776673.49262: checking for max_fail_percentage 9704 1726776673.49263: done checking for max_fail_percentage 9704 1726776673.49264: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.49265: done checking to see if all hosts have failed 9704 1726776673.49265: getting the remaining hosts for this loop 9704 1726776673.49266: done getting the remaining hosts for this loop 9704 1726776673.49270: getting the next task for host managed_node3 9704 1726776673.49276: done getting next task for host managed_node3 9704 1726776673.49279: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables 9704 1726776673.49282: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.49302: getting variables 9704 1726776673.49304: in VariableManager get_vars() 9704 1726776673.49340: Calling all_inventory to load vars for managed_node3 9704 1726776673.49343: Calling groups_inventory to load vars for managed_node3 9704 1726776673.49345: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.49353: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.49356: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.49358: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.49493: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.49654: done with get_vars() 9704 1726776673.49663: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Set version specific variables] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:9 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.040) 0:00:26.426 **** 9704 1726776673.49727: entering _queue_task() for managed_node3/include_tasks 9704 1726776673.49908: worker is 1 (out of 1 available) 9704 1726776673.49924: exiting _queue_task() for managed_node3/include_tasks 9704 1726776673.49937: done queuing things up, now waiting for results queue to drain 9704 1726776673.49940: waiting for pending results... 10623 1726776673.50068: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables 10623 1726776673.50200: in run() - task 120fa90a-8a95-fee7-0cea-0000000006d9 10623 1726776673.50218: variable 'ansible_search_path' from source: unknown 10623 1726776673.50222: variable 'ansible_search_path' from source: unknown 10623 1726776673.50254: calling self._execute() 10623 1726776673.50321: variable 'ansible_host' from source: host vars for 'managed_node3' 10623 1726776673.50331: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10623 1726776673.50340: variable 'omit' from source: magic vars 10623 1726776673.50413: _execute() done 10623 1726776673.50419: dumping result to json 10623 1726776673.50423: done dumping result, returning 10623 1726776673.50431: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set version specific variables [120fa90a-8a95-fee7-0cea-0000000006d9] 10623 1726776673.50438: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006d9 10623 1726776673.50466: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006d9 10623 1726776673.50469: WORKER PROCESS EXITING 9704 1726776673.50574: no more pending results, returning what we have 9704 1726776673.50578: in VariableManager get_vars() 9704 1726776673.50615: Calling all_inventory to load vars for managed_node3 9704 1726776673.50618: Calling groups_inventory to load vars for managed_node3 9704 1726776673.50620: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.50630: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.50633: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.50635: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.50755: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.50876: done with get_vars() 9704 1726776673.50882: variable 'ansible_search_path' from source: unknown 9704 1726776673.50883: variable 'ansible_search_path' from source: unknown 9704 1726776673.50906: we have included files to process 9704 1726776673.50907: generating all_blocks data 9704 1726776673.50907: done generating all_blocks data 9704 1726776673.50913: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml 9704 1726776673.50914: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml 9704 1726776673.50915: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml for managed_node3 9704 1726776673.51374: done processing included file 9704 1726776673.51376: iterating over new_blocks loaded from include file 9704 1726776673.51377: in VariableManager get_vars() 9704 1726776673.51392: done with get_vars() 9704 1726776673.51393: filtering new block on tags 9704 1726776673.51412: done filtering new block on tags 9704 1726776673.51437: in VariableManager get_vars() 9704 1726776673.51454: done with get_vars() 9704 1726776673.51456: filtering new block on tags 9704 1726776673.51480: done filtering new block on tags 9704 1726776673.51482: in VariableManager get_vars() 9704 1726776673.51495: done with get_vars() 9704 1726776673.51496: filtering new block on tags 9704 1726776673.51519: done filtering new block on tags 9704 1726776673.51520: in VariableManager get_vars() 9704 1726776673.51537: done with get_vars() 9704 1726776673.51538: filtering new block on tags 9704 1726776673.51555: done filtering new block on tags 9704 1726776673.51556: done iterating over new_blocks loaded from include file 9704 1726776673.51556: extending task lists for all hosts with included blocks 9704 1726776673.51687: done extending task lists 9704 1726776673.51688: done processing included files 9704 1726776673.51689: results queue empty 9704 1726776673.51689: checking for any_errors_fatal 9704 1726776673.51692: done checking for any_errors_fatal 9704 1726776673.51692: checking for max_fail_percentage 9704 1726776673.51693: done checking for max_fail_percentage 9704 1726776673.51693: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.51694: done checking to see if all hosts have failed 9704 1726776673.51694: getting the remaining hosts for this loop 9704 1726776673.51695: done getting the remaining hosts for this loop 9704 1726776673.51696: getting the next task for host managed_node3 9704 1726776673.51699: done getting next task for host managed_node3 9704 1726776673.51701: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role 9704 1726776673.51703: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.51709: getting variables 9704 1726776673.51709: in VariableManager get_vars() 9704 1726776673.51718: Calling all_inventory to load vars for managed_node3 9704 1726776673.51719: Calling groups_inventory to load vars for managed_node3 9704 1726776673.51720: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.51723: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.51724: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.51726: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.51804: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.51915: done with get_vars() 9704 1726776673.51921: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:2 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.022) 0:00:26.448 **** 9704 1726776673.51970: entering _queue_task() for managed_node3/setup 9704 1726776673.52131: worker is 1 (out of 1 available) 9704 1726776673.52145: exiting _queue_task() for managed_node3/setup 9704 1726776673.52156: done queuing things up, now waiting for results queue to drain 9704 1726776673.52158: waiting for pending results... 10624 1726776673.52283: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role 10624 1726776673.52408: in run() - task 120fa90a-8a95-fee7-0cea-000000000754 10624 1726776673.52424: variable 'ansible_search_path' from source: unknown 10624 1726776673.52430: variable 'ansible_search_path' from source: unknown 10624 1726776673.52458: calling self._execute() 10624 1726776673.52574: variable 'ansible_host' from source: host vars for 'managed_node3' 10624 1726776673.52582: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10624 1726776673.52590: variable 'omit' from source: magic vars 10624 1726776673.52948: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10624 1726776673.54455: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10624 1726776673.54505: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10624 1726776673.54536: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10624 1726776673.54564: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10624 1726776673.54582: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10624 1726776673.54641: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10624 1726776673.54666: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10624 1726776673.54685: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10624 1726776673.54712: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10624 1726776673.54723: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10624 1726776673.54764: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10624 1726776673.54783: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10624 1726776673.54800: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10624 1726776673.54824: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10624 1726776673.54837: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10624 1726776673.54954: variable '__kernel_settings_required_facts' from source: role '' all vars 10624 1726776673.54965: variable 'ansible_facts' from source: unknown 10624 1726776673.55023: Evaluated conditional (__kernel_settings_required_facts | difference(ansible_facts.keys() | list) | length > 0): False 10624 1726776673.55029: when evaluation is False, skipping this task 10624 1726776673.55033: _execute() done 10624 1726776673.55037: dumping result to json 10624 1726776673.55041: done dumping result, returning 10624 1726776673.55050: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure ansible_facts used by role [120fa90a-8a95-fee7-0cea-000000000754] 10624 1726776673.55056: sending task result for task 120fa90a-8a95-fee7-0cea-000000000754 10624 1726776673.55077: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000754 10624 1726776673.55080: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_required_facts | difference(ansible_facts.keys() | list) | length > 0", "skip_reason": "Conditional result was False" } 9704 1726776673.55243: no more pending results, returning what we have 9704 1726776673.55245: results queue empty 9704 1726776673.55246: checking for any_errors_fatal 9704 1726776673.55247: done checking for any_errors_fatal 9704 1726776673.55248: checking for max_fail_percentage 9704 1726776673.55249: done checking for max_fail_percentage 9704 1726776673.55250: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.55250: done checking to see if all hosts have failed 9704 1726776673.55251: getting the remaining hosts for this loop 9704 1726776673.55252: done getting the remaining hosts for this loop 9704 1726776673.55256: getting the next task for host managed_node3 9704 1726776673.55263: done getting next task for host managed_node3 9704 1726776673.55266: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree 9704 1726776673.55270: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.55282: getting variables 9704 1726776673.55283: in VariableManager get_vars() 9704 1726776673.55307: Calling all_inventory to load vars for managed_node3 9704 1726776673.55309: Calling groups_inventory to load vars for managed_node3 9704 1726776673.55311: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.55317: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.55319: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.55321: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.55424: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.55547: done with get_vars() 9704 1726776673.55554: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Check if system is ostree] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:10 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.036) 0:00:26.484 **** 9704 1726776673.55617: entering _queue_task() for managed_node3/stat 9704 1726776673.55767: worker is 1 (out of 1 available) 9704 1726776673.55782: exiting _queue_task() for managed_node3/stat 9704 1726776673.55794: done queuing things up, now waiting for results queue to drain 9704 1726776673.55796: waiting for pending results... 10625 1726776673.55917: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree 10625 1726776673.56038: in run() - task 120fa90a-8a95-fee7-0cea-000000000756 10625 1726776673.56055: variable 'ansible_search_path' from source: unknown 10625 1726776673.56059: variable 'ansible_search_path' from source: unknown 10625 1726776673.56084: calling self._execute() 10625 1726776673.56151: variable 'ansible_host' from source: host vars for 'managed_node3' 10625 1726776673.56159: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10625 1726776673.56167: variable 'omit' from source: magic vars 10625 1726776673.56487: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10625 1726776673.56664: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10625 1726776673.56721: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10625 1726776673.56747: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10625 1726776673.56771: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10625 1726776673.56825: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10625 1726776673.56856: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10625 1726776673.56875: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10625 1726776673.56895: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10625 1726776673.56979: variable '__kernel_settings_is_ostree' from source: set_fact 10625 1726776673.56990: Evaluated conditional (not __kernel_settings_is_ostree is defined): False 10625 1726776673.56995: when evaluation is False, skipping this task 10625 1726776673.56999: _execute() done 10625 1726776673.57003: dumping result to json 10625 1726776673.57007: done dumping result, returning 10625 1726776673.57012: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if system is ostree [120fa90a-8a95-fee7-0cea-000000000756] 10625 1726776673.57017: sending task result for task 120fa90a-8a95-fee7-0cea-000000000756 10625 1726776673.57041: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000756 10625 1726776673.57045: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_ostree is defined", "skip_reason": "Conditional result was False" } 9704 1726776673.57152: no more pending results, returning what we have 9704 1726776673.57155: results queue empty 9704 1726776673.57156: checking for any_errors_fatal 9704 1726776673.57161: done checking for any_errors_fatal 9704 1726776673.57162: checking for max_fail_percentage 9704 1726776673.57164: done checking for max_fail_percentage 9704 1726776673.57164: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.57165: done checking to see if all hosts have failed 9704 1726776673.57165: getting the remaining hosts for this loop 9704 1726776673.57167: done getting the remaining hosts for this loop 9704 1726776673.57169: getting the next task for host managed_node3 9704 1726776673.57176: done getting next task for host managed_node3 9704 1726776673.57179: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree 9704 1726776673.57183: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.57196: getting variables 9704 1726776673.57197: in VariableManager get_vars() 9704 1726776673.57226: Calling all_inventory to load vars for managed_node3 9704 1726776673.57229: Calling groups_inventory to load vars for managed_node3 9704 1726776673.57232: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.57239: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.57240: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.57242: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.57342: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.57487: done with get_vars() 9704 1726776673.57494: done getting variables 9704 1726776673.57532: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:15 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.019) 0:00:26.504 **** 9704 1726776673.57555: entering _queue_task() for managed_node3/set_fact 9704 1726776673.57701: worker is 1 (out of 1 available) 9704 1726776673.57713: exiting _queue_task() for managed_node3/set_fact 9704 1726776673.57725: done queuing things up, now waiting for results queue to drain 9704 1726776673.57727: waiting for pending results... 10626 1726776673.57845: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree 10626 1726776673.57967: in run() - task 120fa90a-8a95-fee7-0cea-000000000757 10626 1726776673.57982: variable 'ansible_search_path' from source: unknown 10626 1726776673.57986: variable 'ansible_search_path' from source: unknown 10626 1726776673.58012: calling self._execute() 10626 1726776673.58079: variable 'ansible_host' from source: host vars for 'managed_node3' 10626 1726776673.58087: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10626 1726776673.58096: variable 'omit' from source: magic vars 10626 1726776673.58410: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10626 1726776673.58584: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10626 1726776673.58619: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10626 1726776673.58648: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10626 1726776673.58674: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10626 1726776673.58734: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10626 1726776673.58757: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10626 1726776673.58776: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10626 1726776673.58794: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10626 1726776673.58878: variable '__kernel_settings_is_ostree' from source: set_fact 10626 1726776673.58889: Evaluated conditional (not __kernel_settings_is_ostree is defined): False 10626 1726776673.58893: when evaluation is False, skipping this task 10626 1726776673.58897: _execute() done 10626 1726776673.58901: dumping result to json 10626 1726776673.58904: done dumping result, returning 10626 1726776673.58910: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate system is ostree [120fa90a-8a95-fee7-0cea-000000000757] 10626 1726776673.58915: sending task result for task 120fa90a-8a95-fee7-0cea-000000000757 10626 1726776673.58940: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000757 10626 1726776673.58945: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_ostree is defined", "skip_reason": "Conditional result was False" } 9704 1726776673.59075: no more pending results, returning what we have 9704 1726776673.59078: results queue empty 9704 1726776673.59079: checking for any_errors_fatal 9704 1726776673.59084: done checking for any_errors_fatal 9704 1726776673.59085: checking for max_fail_percentage 9704 1726776673.59087: done checking for max_fail_percentage 9704 1726776673.59087: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.59088: done checking to see if all hosts have failed 9704 1726776673.59088: getting the remaining hosts for this loop 9704 1726776673.59090: done getting the remaining hosts for this loop 9704 1726776673.59093: getting the next task for host managed_node3 9704 1726776673.59101: done getting next task for host managed_node3 9704 1726776673.59104: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin 9704 1726776673.59108: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.59121: getting variables 9704 1726776673.59122: in VariableManager get_vars() 9704 1726776673.59151: Calling all_inventory to load vars for managed_node3 9704 1726776673.59153: Calling groups_inventory to load vars for managed_node3 9704 1726776673.59155: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.59161: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.59162: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.59164: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.59266: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.59390: done with get_vars() 9704 1726776673.59397: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:22 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.019) 0:00:26.523 **** 9704 1726776673.59464: entering _queue_task() for managed_node3/stat 9704 1726776673.59622: worker is 1 (out of 1 available) 9704 1726776673.59638: exiting _queue_task() for managed_node3/stat 9704 1726776673.59649: done queuing things up, now waiting for results queue to drain 9704 1726776673.59651: waiting for pending results... 10627 1726776673.59779: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin 10627 1726776673.59912: in run() - task 120fa90a-8a95-fee7-0cea-000000000759 10627 1726776673.59930: variable 'ansible_search_path' from source: unknown 10627 1726776673.59934: variable 'ansible_search_path' from source: unknown 10627 1726776673.59964: calling self._execute() 10627 1726776673.60026: variable 'ansible_host' from source: host vars for 'managed_node3' 10627 1726776673.60035: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10627 1726776673.60047: variable 'omit' from source: magic vars 10627 1726776673.60373: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10627 1726776673.60610: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10627 1726776673.60644: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10627 1726776673.60673: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10627 1726776673.60699: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10627 1726776673.60760: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10627 1726776673.60781: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10627 1726776673.60800: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10627 1726776673.60818: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10627 1726776673.60901: variable '__kernel_settings_is_transactional' from source: set_fact 10627 1726776673.60912: Evaluated conditional (not __kernel_settings_is_transactional is defined): False 10627 1726776673.60916: when evaluation is False, skipping this task 10627 1726776673.60920: _execute() done 10627 1726776673.60924: dumping result to json 10627 1726776673.60927: done dumping result, returning 10627 1726776673.60936: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check if transactional-update exists in /sbin [120fa90a-8a95-fee7-0cea-000000000759] 10627 1726776673.60941: sending task result for task 120fa90a-8a95-fee7-0cea-000000000759 10627 1726776673.60966: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000759 10627 1726776673.60968: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_transactional is defined", "skip_reason": "Conditional result was False" } 9704 1726776673.61143: no more pending results, returning what we have 9704 1726776673.61146: results queue empty 9704 1726776673.61147: checking for any_errors_fatal 9704 1726776673.61153: done checking for any_errors_fatal 9704 1726776673.61153: checking for max_fail_percentage 9704 1726776673.61154: done checking for max_fail_percentage 9704 1726776673.61155: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.61155: done checking to see if all hosts have failed 9704 1726776673.61155: getting the remaining hosts for this loop 9704 1726776673.61156: done getting the remaining hosts for this loop 9704 1726776673.61159: getting the next task for host managed_node3 9704 1726776673.61164: done getting next task for host managed_node3 9704 1726776673.61166: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists 9704 1726776673.61169: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.61180: getting variables 9704 1726776673.61181: in VariableManager get_vars() 9704 1726776673.61203: Calling all_inventory to load vars for managed_node3 9704 1726776673.61205: Calling groups_inventory to load vars for managed_node3 9704 1726776673.61206: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.61214: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.61216: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.61218: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.61367: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.61488: done with get_vars() 9704 1726776673.61494: done getting variables 9704 1726776673.61533: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:27 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.020) 0:00:26.544 **** 9704 1726776673.61558: entering _queue_task() for managed_node3/set_fact 9704 1726776673.61732: worker is 1 (out of 1 available) 9704 1726776673.61747: exiting _queue_task() for managed_node3/set_fact 9704 1726776673.61758: done queuing things up, now waiting for results queue to drain 9704 1726776673.61760: waiting for pending results... 10628 1726776673.61880: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists 10628 1726776673.62002: in run() - task 120fa90a-8a95-fee7-0cea-00000000075a 10628 1726776673.62016: variable 'ansible_search_path' from source: unknown 10628 1726776673.62021: variable 'ansible_search_path' from source: unknown 10628 1726776673.62051: calling self._execute() 10628 1726776673.62117: variable 'ansible_host' from source: host vars for 'managed_node3' 10628 1726776673.62125: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10628 1726776673.62135: variable 'omit' from source: magic vars 10628 1726776673.62456: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10628 1726776673.62630: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10628 1726776673.62666: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10628 1726776673.62691: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10628 1726776673.62723: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10628 1726776673.62800: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10628 1726776673.62823: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10628 1726776673.62865: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10628 1726776673.62890: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10628 1726776673.62997: variable '__kernel_settings_is_transactional' from source: set_fact 10628 1726776673.63011: Evaluated conditional (not __kernel_settings_is_transactional is defined): False 10628 1726776673.63016: when evaluation is False, skipping this task 10628 1726776673.63019: _execute() done 10628 1726776673.63022: dumping result to json 10628 1726776673.63025: done dumping result, returning 10628 1726776673.63032: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag if transactional-update exists [120fa90a-8a95-fee7-0cea-00000000075a] 10628 1726776673.63038: sending task result for task 120fa90a-8a95-fee7-0cea-00000000075a 10628 1726776673.63070: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000075a 10628 1726776673.63075: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "not __kernel_settings_is_transactional is defined", "skip_reason": "Conditional result was False" } 9704 1726776673.63403: no more pending results, returning what we have 9704 1726776673.63406: results queue empty 9704 1726776673.63407: checking for any_errors_fatal 9704 1726776673.63414: done checking for any_errors_fatal 9704 1726776673.63414: checking for max_fail_percentage 9704 1726776673.63416: done checking for max_fail_percentage 9704 1726776673.63417: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.63418: done checking to see if all hosts have failed 9704 1726776673.63418: getting the remaining hosts for this loop 9704 1726776673.63419: done getting the remaining hosts for this loop 9704 1726776673.63423: getting the next task for host managed_node3 9704 1726776673.63433: done getting next task for host managed_node3 9704 1726776673.63437: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables 9704 1726776673.63441: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.63460: getting variables 9704 1726776673.63462: in VariableManager get_vars() 9704 1726776673.63494: Calling all_inventory to load vars for managed_node3 9704 1726776673.63497: Calling groups_inventory to load vars for managed_node3 9704 1726776673.63499: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.63507: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.63513: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.63516: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.63632: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.63765: done with get_vars() 9704 1726776673.63772: done getting variables 9704 1726776673.63811: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set platform/version specific variables] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/set_vars.yml:31 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.022) 0:00:26.567 **** 9704 1726776673.63837: entering _queue_task() for managed_node3/include_vars 9704 1726776673.63989: worker is 1 (out of 1 available) 9704 1726776673.64002: exiting _queue_task() for managed_node3/include_vars 9704 1726776673.64014: done queuing things up, now waiting for results queue to drain 9704 1726776673.64016: waiting for pending results... 10630 1726776673.64133: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables 10630 1726776673.64264: in run() - task 120fa90a-8a95-fee7-0cea-00000000075c 10630 1726776673.64279: variable 'ansible_search_path' from source: unknown 10630 1726776673.64283: variable 'ansible_search_path' from source: unknown 10630 1726776673.64309: calling self._execute() 10630 1726776673.64377: variable 'ansible_host' from source: host vars for 'managed_node3' 10630 1726776673.64385: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10630 1726776673.64393: variable 'omit' from source: magic vars 10630 1726776673.64464: variable 'omit' from source: magic vars 10630 1726776673.64508: variable 'omit' from source: magic vars 10630 1726776673.64759: variable 'ffparams' from source: task vars 10630 1726776673.64910: variable 'ansible_facts' from source: unknown 10630 1726776673.65033: variable 'ansible_facts' from source: unknown 10630 1726776673.65123: variable 'ansible_facts' from source: unknown 10630 1726776673.65208: variable 'ansible_facts' from source: unknown 10630 1726776673.65287: variable 'role_path' from source: magic vars 10630 1726776673.65403: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup 10630 1726776673.65566: Loaded config def from plugin (lookup/first_found) 10630 1726776673.65573: Loading LookupModule 'first_found' from /usr/local/lib/python3.12/site-packages/ansible/plugins/lookup/first_found.py 10630 1726776673.65600: variable 'ansible_search_path' from source: unknown 10630 1726776673.65619: variable 'ansible_search_path' from source: unknown 10630 1726776673.65628: variable 'ansible_search_path' from source: unknown 10630 1726776673.65637: variable 'ansible_search_path' from source: unknown 10630 1726776673.65644: variable 'ansible_search_path' from source: unknown 10630 1726776673.65660: variable 'omit' from source: magic vars 10630 1726776673.65678: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10630 1726776673.65696: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10630 1726776673.65711: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10630 1726776673.65726: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10630 1726776673.65737: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10630 1726776673.65760: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10630 1726776673.65764: variable 'ansible_host' from source: host vars for 'managed_node3' 10630 1726776673.65768: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10630 1726776673.65831: Set connection var ansible_module_compression to ZIP_DEFLATED 10630 1726776673.65839: Set connection var ansible_shell_type to sh 10630 1726776673.65849: Set connection var ansible_timeout to 10 10630 1726776673.65855: Set connection var ansible_pipelining to False 10630 1726776673.65862: Set connection var ansible_shell_executable to /bin/sh 10630 1726776673.65865: Set connection var ansible_connection to ssh 10630 1726776673.65880: variable 'ansible_shell_executable' from source: unknown 10630 1726776673.65883: variable 'ansible_connection' from source: unknown 10630 1726776673.65886: variable 'ansible_module_compression' from source: unknown 10630 1726776673.65890: variable 'ansible_shell_type' from source: unknown 10630 1726776673.65893: variable 'ansible_shell_executable' from source: unknown 10630 1726776673.65896: variable 'ansible_host' from source: host vars for 'managed_node3' 10630 1726776673.65900: variable 'ansible_pipelining' from source: unknown 10630 1726776673.65904: variable 'ansible_timeout' from source: unknown 10630 1726776673.65908: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10630 1726776673.65998: Loading ActionModule 'include_vars' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/include_vars.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10630 1726776673.66010: variable 'omit' from source: magic vars 10630 1726776673.66016: starting attempt loop 10630 1726776673.66019: running the handler 10630 1726776673.66077: handler run complete 10630 1726776673.66087: attempt loop complete, returning result 10630 1726776673.66091: _execute() done 10630 1726776673.66095: dumping result to json 10630 1726776673.66099: done dumping result, returning 10630 1726776673.66104: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set platform/version specific variables [120fa90a-8a95-fee7-0cea-00000000075c] 10630 1726776673.66111: sending task result for task 120fa90a-8a95-fee7-0cea-00000000075c 10630 1726776673.66162: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000075c 10630 1726776673.66166: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_packages": [ "tuned", "python3-configobj" ], "__kernel_settings_services": [ "tuned" ] }, "ansible_included_var_files": [ "/tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/vars/default.yml" ], "changed": false } 9704 1726776673.66496: no more pending results, returning what we have 9704 1726776673.66499: results queue empty 9704 1726776673.66500: checking for any_errors_fatal 9704 1726776673.66504: done checking for any_errors_fatal 9704 1726776673.66505: checking for max_fail_percentage 9704 1726776673.66506: done checking for max_fail_percentage 9704 1726776673.66507: checking to see if all hosts have failed and the running result is not ok 9704 1726776673.66507: done checking to see if all hosts have failed 9704 1726776673.66508: getting the remaining hosts for this loop 9704 1726776673.66509: done getting the remaining hosts for this loop 9704 1726776673.66512: getting the next task for host managed_node3 9704 1726776673.66520: done getting next task for host managed_node3 9704 1726776673.66524: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed 9704 1726776673.66527: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776673.66540: getting variables 9704 1726776673.66541: in VariableManager get_vars() 9704 1726776673.66575: Calling all_inventory to load vars for managed_node3 9704 1726776673.66578: Calling groups_inventory to load vars for managed_node3 9704 1726776673.66580: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776673.66593: Calling all_plugins_play to load vars for managed_node3 9704 1726776673.66596: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776673.66598: Calling groups_plugins_play to load vars for managed_node3 9704 1726776673.66816: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776673.67021: done with get_vars() 9704 1726776673.67031: done getting variables 9704 1726776673.67086: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure required packages are installed] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:12 Thursday 19 September 2024 16:11:13 -0400 (0:00:00.032) 0:00:26.599 **** 9704 1726776673.67115: entering _queue_task() for managed_node3/package 9704 1726776673.67307: worker is 1 (out of 1 available) 9704 1726776673.67322: exiting _queue_task() for managed_node3/package 9704 1726776673.67337: done queuing things up, now waiting for results queue to drain 9704 1726776673.67339: waiting for pending results... 10632 1726776673.67577: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed 10632 1726776673.67728: in run() - task 120fa90a-8a95-fee7-0cea-0000000006da 10632 1726776673.67750: variable 'ansible_search_path' from source: unknown 10632 1726776673.67755: variable 'ansible_search_path' from source: unknown 10632 1726776673.67790: calling self._execute() 10632 1726776673.67882: variable 'ansible_host' from source: host vars for 'managed_node3' 10632 1726776673.67891: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10632 1726776673.67900: variable 'omit' from source: magic vars 10632 1726776673.67998: variable 'omit' from source: magic vars 10632 1726776673.68056: variable 'omit' from source: magic vars 10632 1726776673.68085: variable '__kernel_settings_packages' from source: include_vars 10632 1726776673.68368: variable '__kernel_settings_packages' from source: include_vars 10632 1726776673.68637: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10632 1726776673.70521: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10632 1726776673.70577: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10632 1726776673.70606: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10632 1726776673.70635: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10632 1726776673.70659: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10632 1726776673.70721: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10632 1726776673.70748: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10632 1726776673.70767: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10632 1726776673.70794: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10632 1726776673.70805: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10632 1726776673.70879: variable '__kernel_settings_is_ostree' from source: set_fact 10632 1726776673.70885: variable 'omit' from source: magic vars 10632 1726776673.70908: variable 'omit' from source: magic vars 10632 1726776673.70927: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10632 1726776673.70951: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10632 1726776673.70968: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10632 1726776673.70981: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10632 1726776673.70990: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10632 1726776673.71013: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10632 1726776673.71018: variable 'ansible_host' from source: host vars for 'managed_node3' 10632 1726776673.71022: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10632 1726776673.71090: Set connection var ansible_module_compression to ZIP_DEFLATED 10632 1726776673.71097: Set connection var ansible_shell_type to sh 10632 1726776673.71106: Set connection var ansible_timeout to 10 10632 1726776673.71111: Set connection var ansible_pipelining to False 10632 1726776673.71118: Set connection var ansible_shell_executable to /bin/sh 10632 1726776673.71121: Set connection var ansible_connection to ssh 10632 1726776673.71138: variable 'ansible_shell_executable' from source: unknown 10632 1726776673.71142: variable 'ansible_connection' from source: unknown 10632 1726776673.71145: variable 'ansible_module_compression' from source: unknown 10632 1726776673.71151: variable 'ansible_shell_type' from source: unknown 10632 1726776673.71154: variable 'ansible_shell_executable' from source: unknown 10632 1726776673.71157: variable 'ansible_host' from source: host vars for 'managed_node3' 10632 1726776673.71161: variable 'ansible_pipelining' from source: unknown 10632 1726776673.71164: variable 'ansible_timeout' from source: unknown 10632 1726776673.71168: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10632 1726776673.71227: Loading ActionModule 'package' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/package.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10632 1726776673.71240: variable 'omit' from source: magic vars 10632 1726776673.71248: starting attempt loop 10632 1726776673.71252: running the handler 10632 1726776673.71309: variable 'ansible_facts' from source: unknown 10632 1726776673.71388: _low_level_execute_command(): starting 10632 1726776673.71397: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10632 1726776673.73659: stdout chunk (state=2): >>>/root <<< 10632 1726776673.73778: stderr chunk (state=3): >>><<< 10632 1726776673.73785: stdout chunk (state=3): >>><<< 10632 1726776673.73800: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10632 1726776673.73812: _low_level_execute_command(): starting 10632 1726776673.73817: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016 `" && echo ansible-tmp-1726776673.738072-10632-199180718653016="` echo /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016 `" ) && sleep 0' 10632 1726776673.76229: stdout chunk (state=2): >>>ansible-tmp-1726776673.738072-10632-199180718653016=/root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016 <<< 10632 1726776673.76356: stderr chunk (state=3): >>><<< 10632 1726776673.76363: stdout chunk (state=3): >>><<< 10632 1726776673.76376: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776673.738072-10632-199180718653016=/root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016 , stderr= 10632 1726776673.76397: variable 'ansible_module_compression' from source: unknown 10632 1726776673.76438: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.dnf-ZIP_DEFLATED 10632 1726776673.76470: variable 'ansible_facts' from source: unknown 10632 1726776673.76557: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/AnsiballZ_dnf.py 10632 1726776673.76646: Sending initial data 10632 1726776673.76653: Sent initial data (150 bytes) 10632 1726776673.79126: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpjt7l4hfn /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/AnsiballZ_dnf.py <<< 10632 1726776673.80421: stderr chunk (state=3): >>><<< 10632 1726776673.80429: stdout chunk (state=3): >>><<< 10632 1726776673.80447: done transferring module to remote 10632 1726776673.80460: _low_level_execute_command(): starting 10632 1726776673.80465: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/ /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/AnsiballZ_dnf.py && sleep 0' 10632 1726776673.82815: stderr chunk (state=2): >>><<< 10632 1726776673.82827: stdout chunk (state=2): >>><<< 10632 1726776673.82844: _low_level_execute_command() done: rc=0, stdout=, stderr= 10632 1726776673.82850: _low_level_execute_command(): starting 10632 1726776673.82856: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/AnsiballZ_dnf.py && sleep 0' 10632 1726776676.39941: stdout chunk (state=2): >>> {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["tuned", "python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} <<< 10632 1726776676.47279: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10632 1726776676.47329: stderr chunk (state=3): >>><<< 10632 1726776676.47338: stdout chunk (state=3): >>><<< 10632 1726776676.47357: _low_level_execute_command() done: rc=0, stdout= {"msg": "Nothing to do", "changed": false, "results": [], "rc": 0, "invocation": {"module_args": {"name": ["tuned", "python3-configobj"], "state": "present", "allow_downgrade": false, "autoremove": false, "bugfix": false, "cacheonly": false, "disable_gpg_check": false, "disable_plugin": [], "disablerepo": [], "download_only": false, "enable_plugin": [], "enablerepo": [], "exclude": [], "installroot": "/", "install_repoquery": true, "install_weak_deps": true, "security": false, "skip_broken": false, "update_cache": false, "update_only": false, "validate_certs": true, "sslverify": true, "lock_timeout": 30, "allowerasing": false, "nobest": false, "use_backend": "auto", "conf_file": null, "disable_excludes": null, "download_dir": null, "list": null, "releasever": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10632 1726776676.47393: done with _execute_module (ansible.legacy.dnf, {'name': ['tuned', 'python3-configobj'], 'state': 'present', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.dnf', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10632 1726776676.47401: _low_level_execute_command(): starting 10632 1726776676.47408: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776673.738072-10632-199180718653016/ > /dev/null 2>&1 && sleep 0' 10632 1726776676.49780: stderr chunk (state=2): >>><<< 10632 1726776676.49790: stdout chunk (state=2): >>><<< 10632 1726776676.49806: _low_level_execute_command() done: rc=0, stdout=, stderr= 10632 1726776676.49814: handler run complete 10632 1726776676.49842: attempt loop complete, returning result 10632 1726776676.49847: _execute() done 10632 1726776676.49850: dumping result to json 10632 1726776676.49856: done dumping result, returning 10632 1726776676.49864: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required packages are installed [120fa90a-8a95-fee7-0cea-0000000006da] 10632 1726776676.49870: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006da 10632 1726776676.49899: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006da 10632 1726776676.49902: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do 9704 1726776676.50074: no more pending results, returning what we have 9704 1726776676.50077: results queue empty 9704 1726776676.50078: checking for any_errors_fatal 9704 1726776676.50086: done checking for any_errors_fatal 9704 1726776676.50086: checking for max_fail_percentage 9704 1726776676.50088: done checking for max_fail_percentage 9704 1726776676.50088: checking to see if all hosts have failed and the running result is not ok 9704 1726776676.50089: done checking to see if all hosts have failed 9704 1726776676.50089: getting the remaining hosts for this loop 9704 1726776676.50091: done getting the remaining hosts for this loop 9704 1726776676.50094: getting the next task for host managed_node3 9704 1726776676.50102: done getting next task for host managed_node3 9704 1726776676.50106: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes 9704 1726776676.50109: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776676.50118: getting variables 9704 1726776676.50120: in VariableManager get_vars() 9704 1726776676.50157: Calling all_inventory to load vars for managed_node3 9704 1726776676.50160: Calling groups_inventory to load vars for managed_node3 9704 1726776676.50161: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776676.50170: Calling all_plugins_play to load vars for managed_node3 9704 1726776676.50172: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776676.50175: Calling groups_plugins_play to load vars for managed_node3 9704 1726776676.50288: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776676.50417: done with get_vars() 9704 1726776676.50425: done getting variables 9704 1726776676.50472: Loading ActionModule 'debug' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/debug.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:24 Thursday 19 September 2024 16:11:16 -0400 (0:00:02.833) 0:00:29.433 **** 9704 1726776676.50499: entering _queue_task() for managed_node3/debug 9704 1726776676.50673: worker is 1 (out of 1 available) 9704 1726776676.50688: exiting _queue_task() for managed_node3/debug 9704 1726776676.50699: done queuing things up, now waiting for results queue to drain 9704 1726776676.50702: waiting for pending results... 10691 1726776676.50824: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes 10691 1726776676.50948: in run() - task 120fa90a-8a95-fee7-0cea-0000000006dc 10691 1726776676.50967: variable 'ansible_search_path' from source: unknown 10691 1726776676.50971: variable 'ansible_search_path' from source: unknown 10691 1726776676.51001: calling self._execute() 10691 1726776676.51074: variable 'ansible_host' from source: host vars for 'managed_node3' 10691 1726776676.51083: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10691 1726776676.51091: variable 'omit' from source: magic vars 10691 1726776676.51423: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10691 1726776676.52974: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10691 1726776676.53018: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10691 1726776676.53045: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10691 1726776676.53071: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10691 1726776676.53090: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10691 1726776676.53143: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10691 1726776676.53173: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10691 1726776676.53189: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10691 1726776676.53214: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10691 1726776676.53223: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10691 1726776676.53297: variable '__kernel_settings_is_transactional' from source: set_fact 10691 1726776676.53311: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 10691 1726776676.53315: when evaluation is False, skipping this task 10691 1726776676.53317: _execute() done 10691 1726776676.53319: dumping result to json 10691 1726776676.53321: done dumping result, returning 10691 1726776676.53325: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Notify user that reboot is needed to apply changes [120fa90a-8a95-fee7-0cea-0000000006dc] 10691 1726776676.53330: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006dc 10691 1726776676.53361: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006dc 10691 1726776676.53365: WORKER PROCESS EXITING skipping: [managed_node3] => { "false_condition": "__kernel_settings_is_transactional | d(false)" } 9704 1726776676.53468: no more pending results, returning what we have 9704 1726776676.53471: results queue empty 9704 1726776676.53472: checking for any_errors_fatal 9704 1726776676.53480: done checking for any_errors_fatal 9704 1726776676.53480: checking for max_fail_percentage 9704 1726776676.53482: done checking for max_fail_percentage 9704 1726776676.53482: checking to see if all hosts have failed and the running result is not ok 9704 1726776676.53483: done checking to see if all hosts have failed 9704 1726776676.53483: getting the remaining hosts for this loop 9704 1726776676.53485: done getting the remaining hosts for this loop 9704 1726776676.53488: getting the next task for host managed_node3 9704 1726776676.53494: done getting next task for host managed_node3 9704 1726776676.53497: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems 9704 1726776676.53500: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776676.53513: getting variables 9704 1726776676.53515: in VariableManager get_vars() 9704 1726776676.53549: Calling all_inventory to load vars for managed_node3 9704 1726776676.53551: Calling groups_inventory to load vars for managed_node3 9704 1726776676.53553: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776676.53561: Calling all_plugins_play to load vars for managed_node3 9704 1726776676.53564: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776676.53566: Calling groups_plugins_play to load vars for managed_node3 9704 1726776676.53677: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776676.53843: done with get_vars() 9704 1726776676.53849: done getting variables 9704 1726776676.53892: Loading ActionModule 'reboot' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/reboot.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Reboot transactional update systems] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:29 Thursday 19 September 2024 16:11:16 -0400 (0:00:00.034) 0:00:29.467 **** 9704 1726776676.53913: entering _queue_task() for managed_node3/reboot 9704 1726776676.54070: worker is 1 (out of 1 available) 9704 1726776676.54086: exiting _queue_task() for managed_node3/reboot 9704 1726776676.54099: done queuing things up, now waiting for results queue to drain 9704 1726776676.54101: waiting for pending results... 10692 1726776676.54226: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems 10692 1726776676.54343: in run() - task 120fa90a-8a95-fee7-0cea-0000000006dd 10692 1726776676.54363: variable 'ansible_search_path' from source: unknown 10692 1726776676.54367: variable 'ansible_search_path' from source: unknown 10692 1726776676.54394: calling self._execute() 10692 1726776676.54467: variable 'ansible_host' from source: host vars for 'managed_node3' 10692 1726776676.54475: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10692 1726776676.54484: variable 'omit' from source: magic vars 10692 1726776676.54810: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10692 1726776676.56294: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10692 1726776676.56348: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10692 1726776676.56379: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10692 1726776676.56406: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10692 1726776676.56427: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10692 1726776676.56482: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10692 1726776676.56502: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10692 1726776676.56520: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10692 1726776676.56556: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10692 1726776676.56568: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10692 1726776676.56646: variable '__kernel_settings_is_transactional' from source: set_fact 10692 1726776676.56664: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 10692 1726776676.56669: when evaluation is False, skipping this task 10692 1726776676.56672: _execute() done 10692 1726776676.56676: dumping result to json 10692 1726776676.56680: done dumping result, returning 10692 1726776676.56686: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Reboot transactional update systems [120fa90a-8a95-fee7-0cea-0000000006dd] 10692 1726776676.56692: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006dd 10692 1726776676.56716: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006dd 10692 1726776676.56720: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_is_transactional | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776676.56847: no more pending results, returning what we have 9704 1726776676.56851: results queue empty 9704 1726776676.56851: checking for any_errors_fatal 9704 1726776676.56856: done checking for any_errors_fatal 9704 1726776676.56857: checking for max_fail_percentage 9704 1726776676.56858: done checking for max_fail_percentage 9704 1726776676.56859: checking to see if all hosts have failed and the running result is not ok 9704 1726776676.56859: done checking to see if all hosts have failed 9704 1726776676.56860: getting the remaining hosts for this loop 9704 1726776676.56861: done getting the remaining hosts for this loop 9704 1726776676.56864: getting the next task for host managed_node3 9704 1726776676.56870: done getting next task for host managed_node3 9704 1726776676.56873: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set 9704 1726776676.56876: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776676.56890: getting variables 9704 1726776676.56892: in VariableManager get_vars() 9704 1726776676.56926: Calling all_inventory to load vars for managed_node3 9704 1726776676.56930: Calling groups_inventory to load vars for managed_node3 9704 1726776676.56932: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776676.56940: Calling all_plugins_play to load vars for managed_node3 9704 1726776676.56942: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776676.56945: Calling groups_plugins_play to load vars for managed_node3 9704 1726776676.57056: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776676.57182: done with get_vars() 9704 1726776676.57190: done getting variables 9704 1726776676.57231: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:34 Thursday 19 September 2024 16:11:16 -0400 (0:00:00.033) 0:00:29.501 **** 9704 1726776676.57255: entering _queue_task() for managed_node3/fail 9704 1726776676.57421: worker is 1 (out of 1 available) 9704 1726776676.57438: exiting _queue_task() for managed_node3/fail 9704 1726776676.57450: done queuing things up, now waiting for results queue to drain 9704 1726776676.57452: waiting for pending results... 10693 1726776676.57579: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set 10693 1726776676.57705: in run() - task 120fa90a-8a95-fee7-0cea-0000000006de 10693 1726776676.57722: variable 'ansible_search_path' from source: unknown 10693 1726776676.57726: variable 'ansible_search_path' from source: unknown 10693 1726776676.57758: calling self._execute() 10693 1726776676.57829: variable 'ansible_host' from source: host vars for 'managed_node3' 10693 1726776676.57838: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10693 1726776676.57847: variable 'omit' from source: magic vars 10693 1726776676.58189: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10693 1726776676.59740: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10693 1726776676.59786: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10693 1726776676.59815: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10693 1726776676.59844: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10693 1726776676.59866: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10693 1726776676.59920: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10693 1726776676.60163: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10693 1726776676.60183: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10693 1726776676.60211: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10693 1726776676.60224: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10693 1726776676.60305: variable '__kernel_settings_is_transactional' from source: set_fact 10693 1726776676.60320: Evaluated conditional (__kernel_settings_is_transactional | d(false)): False 10693 1726776676.60325: when evaluation is False, skipping this task 10693 1726776676.60330: _execute() done 10693 1726776676.60334: dumping result to json 10693 1726776676.60338: done dumping result, returning 10693 1726776676.60344: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Fail if reboot is needed and not set [120fa90a-8a95-fee7-0cea-0000000006de] 10693 1726776676.60349: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006de 10693 1726776676.60376: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006de 10693 1726776676.60380: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_is_transactional | d(false)", "skip_reason": "Conditional result was False" } 9704 1726776676.60514: no more pending results, returning what we have 9704 1726776676.60518: results queue empty 9704 1726776676.60518: checking for any_errors_fatal 9704 1726776676.60524: done checking for any_errors_fatal 9704 1726776676.60524: checking for max_fail_percentage 9704 1726776676.60526: done checking for max_fail_percentage 9704 1726776676.60527: checking to see if all hosts have failed and the running result is not ok 9704 1726776676.60527: done checking to see if all hosts have failed 9704 1726776676.60530: getting the remaining hosts for this loop 9704 1726776676.60531: done getting the remaining hosts for this loop 9704 1726776676.60534: getting the next task for host managed_node3 9704 1726776676.60543: done getting next task for host managed_node3 9704 1726776676.60547: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config 9704 1726776676.60550: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776676.60565: getting variables 9704 1726776676.60567: in VariableManager get_vars() 9704 1726776676.60600: Calling all_inventory to load vars for managed_node3 9704 1726776676.60603: Calling groups_inventory to load vars for managed_node3 9704 1726776676.60604: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776676.60613: Calling all_plugins_play to load vars for managed_node3 9704 1726776676.60615: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776676.60618: Calling groups_plugins_play to load vars for managed_node3 9704 1726776676.60773: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776676.60894: done with get_vars() 9704 1726776676.60901: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Read tuned main config] ****** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:42 Thursday 19 September 2024 16:11:16 -0400 (0:00:00.037) 0:00:29.538 **** 9704 1726776676.60961: entering _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776676.61126: worker is 1 (out of 1 available) 9704 1726776676.61142: exiting _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776676.61153: done queuing things up, now waiting for results queue to drain 9704 1726776676.61156: waiting for pending results... 10694 1726776676.61285: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config 10694 1726776676.61397: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e0 10694 1726776676.61414: variable 'ansible_search_path' from source: unknown 10694 1726776676.61418: variable 'ansible_search_path' from source: unknown 10694 1726776676.61446: calling self._execute() 10694 1726776676.61515: variable 'ansible_host' from source: host vars for 'managed_node3' 10694 1726776676.61523: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10694 1726776676.61533: variable 'omit' from source: magic vars 10694 1726776676.61605: variable 'omit' from source: magic vars 10694 1726776676.61647: variable 'omit' from source: magic vars 10694 1726776676.61671: variable '__kernel_settings_tuned_main_conf_file' from source: role '' all vars 10694 1726776676.61878: variable '__kernel_settings_tuned_main_conf_file' from source: role '' all vars 10694 1726776676.61937: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10694 1726776676.61967: variable 'omit' from source: magic vars 10694 1726776676.61996: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10694 1726776676.62021: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10694 1726776676.62042: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10694 1726776676.62058: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10694 1726776676.62068: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10694 1726776676.62091: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10694 1726776676.62096: variable 'ansible_host' from source: host vars for 'managed_node3' 10694 1726776676.62100: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10694 1726776676.62170: Set connection var ansible_module_compression to ZIP_DEFLATED 10694 1726776676.62177: Set connection var ansible_shell_type to sh 10694 1726776676.62185: Set connection var ansible_timeout to 10 10694 1726776676.62191: Set connection var ansible_pipelining to False 10694 1726776676.62198: Set connection var ansible_shell_executable to /bin/sh 10694 1726776676.62201: Set connection var ansible_connection to ssh 10694 1726776676.62217: variable 'ansible_shell_executable' from source: unknown 10694 1726776676.62221: variable 'ansible_connection' from source: unknown 10694 1726776676.62224: variable 'ansible_module_compression' from source: unknown 10694 1726776676.62228: variable 'ansible_shell_type' from source: unknown 10694 1726776676.62233: variable 'ansible_shell_executable' from source: unknown 10694 1726776676.62236: variable 'ansible_host' from source: host vars for 'managed_node3' 10694 1726776676.62240: variable 'ansible_pipelining' from source: unknown 10694 1726776676.62244: variable 'ansible_timeout' from source: unknown 10694 1726776676.62248: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10694 1726776676.62371: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10694 1726776676.62382: variable 'omit' from source: magic vars 10694 1726776676.62388: starting attempt loop 10694 1726776676.62391: running the handler 10694 1726776676.62403: _low_level_execute_command(): starting 10694 1726776676.62410: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10694 1726776676.64688: stdout chunk (state=2): >>>/root <<< 10694 1726776676.64809: stderr chunk (state=3): >>><<< 10694 1726776676.64815: stdout chunk (state=3): >>><<< 10694 1726776676.64833: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10694 1726776676.64846: _low_level_execute_command(): starting 10694 1726776676.64852: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391 `" && echo ansible-tmp-1726776676.6484087-10694-260409904361391="` echo /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391 `" ) && sleep 0' 10694 1726776676.67215: stdout chunk (state=2): >>>ansible-tmp-1726776676.6484087-10694-260409904361391=/root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391 <<< 10694 1726776676.67333: stderr chunk (state=3): >>><<< 10694 1726776676.67339: stdout chunk (state=3): >>><<< 10694 1726776676.67354: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776676.6484087-10694-260409904361391=/root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391 , stderr= 10694 1726776676.67387: variable 'ansible_module_compression' from source: unknown 10694 1726776676.67418: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.kernel_settings_get_config-ZIP_DEFLATED 10694 1726776676.67445: variable 'ansible_facts' from source: unknown 10694 1726776676.67510: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/AnsiballZ_kernel_settings_get_config.py 10694 1726776676.67606: Sending initial data 10694 1726776676.67613: Sent initial data (174 bytes) 10694 1726776676.69993: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp6ix7nfk_ /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/AnsiballZ_kernel_settings_get_config.py <<< 10694 1726776676.70943: stderr chunk (state=3): >>><<< 10694 1726776676.70950: stdout chunk (state=3): >>><<< 10694 1726776676.70970: done transferring module to remote 10694 1726776676.70981: _low_level_execute_command(): starting 10694 1726776676.70986: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/ /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10694 1726776676.73228: stderr chunk (state=2): >>><<< 10694 1726776676.73236: stdout chunk (state=2): >>><<< 10694 1726776676.73257: _low_level_execute_command() done: rc=0, stdout=, stderr= 10694 1726776676.73263: _low_level_execute_command(): starting 10694 1726776676.73269: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10694 1726776676.88306: stdout chunk (state=2): >>> {"changed": false, "data": {"daemon": "1", "dynamic_tuning": "0", "sleep_interval": "1", "update_interval": "10", "recommend_command": "1", "reapply_sysctl": "1", "default_instance_priority": "0", "udev_buffer_size": "1MB", "log_file_count": "2", "log_file_max_size": "1MB"}, "invocation": {"module_args": {"path": "/etc/tuned/tuned-main.conf"}}} <<< 10694 1726776676.89266: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10694 1726776676.89310: stderr chunk (state=3): >>><<< 10694 1726776676.89319: stdout chunk (state=3): >>><<< 10694 1726776676.89336: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "data": {"daemon": "1", "dynamic_tuning": "0", "sleep_interval": "1", "update_interval": "10", "recommend_command": "1", "reapply_sysctl": "1", "default_instance_priority": "0", "udev_buffer_size": "1MB", "log_file_count": "2", "log_file_max_size": "1MB"}, "invocation": {"module_args": {"path": "/etc/tuned/tuned-main.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 10694 1726776676.89368: done with _execute_module (fedora.linux_system_roles.kernel_settings_get_config, {'path': '/etc/tuned/tuned-main.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'fedora.linux_system_roles.kernel_settings_get_config', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10694 1726776676.89379: _low_level_execute_command(): starting 10694 1726776676.89385: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776676.6484087-10694-260409904361391/ > /dev/null 2>&1 && sleep 0' 10694 1726776676.91746: stderr chunk (state=2): >>><<< 10694 1726776676.91757: stdout chunk (state=2): >>><<< 10694 1726776676.91772: _low_level_execute_command() done: rc=0, stdout=, stderr= 10694 1726776676.91778: handler run complete 10694 1726776676.91793: attempt loop complete, returning result 10694 1726776676.91797: _execute() done 10694 1726776676.91800: dumping result to json 10694 1726776676.91805: done dumping result, returning 10694 1726776676.91812: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Read tuned main config [120fa90a-8a95-fee7-0cea-0000000006e0] 10694 1726776676.91820: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e0 10694 1726776676.91848: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e0 10694 1726776676.91850: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "data": { "daemon": "1", "default_instance_priority": "0", "dynamic_tuning": "0", "log_file_count": "2", "log_file_max_size": "1MB", "reapply_sysctl": "1", "recommend_command": "1", "sleep_interval": "1", "udev_buffer_size": "1MB", "update_interval": "10" } } 9704 1726776676.92082: no more pending results, returning what we have 9704 1726776676.92085: results queue empty 9704 1726776676.92085: checking for any_errors_fatal 9704 1726776676.92089: done checking for any_errors_fatal 9704 1726776676.92090: checking for max_fail_percentage 9704 1726776676.92091: done checking for max_fail_percentage 9704 1726776676.92091: checking to see if all hosts have failed and the running result is not ok 9704 1726776676.92092: done checking to see if all hosts have failed 9704 1726776676.92092: getting the remaining hosts for this loop 9704 1726776676.92093: done getting the remaining hosts for this loop 9704 1726776676.92095: getting the next task for host managed_node3 9704 1726776676.92100: done getting next task for host managed_node3 9704 1726776676.92103: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory 9704 1726776676.92105: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776676.92112: getting variables 9704 1726776676.92113: in VariableManager get_vars() 9704 1726776676.92141: Calling all_inventory to load vars for managed_node3 9704 1726776676.92143: Calling groups_inventory to load vars for managed_node3 9704 1726776676.92145: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776676.92151: Calling all_plugins_play to load vars for managed_node3 9704 1726776676.92154: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776676.92155: Calling groups_plugins_play to load vars for managed_node3 9704 1726776676.92264: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776676.92390: done with get_vars() 9704 1726776676.92399: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:50 Thursday 19 September 2024 16:11:16 -0400 (0:00:00.315) 0:00:29.853 **** 9704 1726776676.92468: entering _queue_task() for managed_node3/stat 9704 1726776676.92628: worker is 1 (out of 1 available) 9704 1726776676.92644: exiting _queue_task() for managed_node3/stat 9704 1726776676.92656: done queuing things up, now waiting for results queue to drain 9704 1726776676.92658: waiting for pending results... 10710 1726776676.92794: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory 10710 1726776676.92927: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e1 10710 1726776676.92945: variable 'ansible_search_path' from source: unknown 10710 1726776676.92948: variable 'ansible_search_path' from source: unknown 10710 1726776676.92984: variable '__prof_from_conf' from source: task vars 10710 1726776676.93237: variable '__prof_from_conf' from source: task vars 10710 1726776676.93433: variable '__data' from source: task vars 10710 1726776676.93505: variable '__kernel_settings_register_tuned_main' from source: set_fact 10710 1726776676.93664: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10710 1726776676.93674: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10710 1726776676.93715: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10710 1726776676.94095: variable 'omit' from source: magic vars 10710 1726776676.94193: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776676.94203: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776676.94213: variable 'omit' from source: magic vars 10710 1726776676.94459: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10710 1726776676.96187: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10710 1726776676.96234: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10710 1726776676.96264: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10710 1726776676.96301: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10710 1726776676.96323: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10710 1726776676.96378: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10710 1726776676.96399: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10710 1726776676.96418: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10710 1726776676.96448: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10710 1726776676.96462: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10710 1726776676.96535: variable 'item' from source: unknown 10710 1726776676.96554: Evaluated conditional (item | length > 0): False 10710 1726776676.96559: when evaluation is False, skipping this task 10710 1726776676.96585: variable 'item' from source: unknown 10710 1726776676.96635: variable 'item' from source: unknown skipping: [managed_node3] => (item=) => { "ansible_loop_var": "item", "changed": false, "false_condition": "item | length > 0", "item": "", "skip_reason": "Conditional result was False" } 10710 1726776676.96711: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776676.96721: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776676.96732: variable 'omit' from source: magic vars 10710 1726776676.96851: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10710 1726776676.96871: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10710 1726776676.96889: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10710 1726776676.96917: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10710 1726776676.96931: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10710 1726776676.96984: variable 'item' from source: unknown 10710 1726776676.96993: Evaluated conditional (item | length > 0): True 10710 1726776676.96999: variable 'omit' from source: magic vars 10710 1726776676.97037: variable 'omit' from source: magic vars 10710 1726776676.97081: variable 'item' from source: unknown 10710 1726776676.97136: variable 'item' from source: unknown 10710 1726776676.97147: variable 'omit' from source: magic vars 10710 1726776676.97166: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10710 1726776676.97192: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10710 1726776676.97208: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10710 1726776676.97223: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10710 1726776676.97234: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10710 1726776676.97262: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10710 1726776676.97267: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776676.97270: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776676.97366: Set connection var ansible_module_compression to ZIP_DEFLATED 10710 1726776676.97374: Set connection var ansible_shell_type to sh 10710 1726776676.97382: Set connection var ansible_timeout to 10 10710 1726776676.97387: Set connection var ansible_pipelining to False 10710 1726776676.97394: Set connection var ansible_shell_executable to /bin/sh 10710 1726776676.97397: Set connection var ansible_connection to ssh 10710 1726776676.97417: variable 'ansible_shell_executable' from source: unknown 10710 1726776676.97424: variable 'ansible_connection' from source: unknown 10710 1726776676.97427: variable 'ansible_module_compression' from source: unknown 10710 1726776676.97431: variable 'ansible_shell_type' from source: unknown 10710 1726776676.97434: variable 'ansible_shell_executable' from source: unknown 10710 1726776676.97437: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776676.97441: variable 'ansible_pipelining' from source: unknown 10710 1726776676.97444: variable 'ansible_timeout' from source: unknown 10710 1726776676.97448: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776676.97577: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10710 1726776676.97588: variable 'omit' from source: magic vars 10710 1726776676.97593: starting attempt loop 10710 1726776676.97596: running the handler 10710 1726776676.97610: _low_level_execute_command(): starting 10710 1726776676.97620: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10710 1726776676.99998: stdout chunk (state=2): >>>/root <<< 10710 1726776677.00114: stderr chunk (state=3): >>><<< 10710 1726776677.00122: stdout chunk (state=3): >>><<< 10710 1726776677.00141: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10710 1726776677.00154: _low_level_execute_command(): starting 10710 1726776677.00161: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640 `" && echo ansible-tmp-1726776677.0014873-10710-274625893143640="` echo /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640 `" ) && sleep 0' 10710 1726776677.02522: stdout chunk (state=2): >>>ansible-tmp-1726776677.0014873-10710-274625893143640=/root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640 <<< 10710 1726776677.02644: stderr chunk (state=3): >>><<< 10710 1726776677.02650: stdout chunk (state=3): >>><<< 10710 1726776677.02665: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776677.0014873-10710-274625893143640=/root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640 , stderr= 10710 1726776677.02700: variable 'ansible_module_compression' from source: unknown 10710 1726776677.02742: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10710 1726776677.02771: variable 'ansible_facts' from source: unknown 10710 1726776677.02837: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/AnsiballZ_stat.py 10710 1726776677.02933: Sending initial data 10710 1726776677.02941: Sent initial data (152 bytes) 10710 1726776677.05358: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpghx_ofgc /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/AnsiballZ_stat.py <<< 10710 1726776677.06317: stderr chunk (state=3): >>><<< 10710 1726776677.06325: stdout chunk (state=3): >>><<< 10710 1726776677.06345: done transferring module to remote 10710 1726776677.06357: _low_level_execute_command(): starting 10710 1726776677.06363: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/ /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/AnsiballZ_stat.py && sleep 0' 10710 1726776677.08637: stderr chunk (state=2): >>><<< 10710 1726776677.08645: stdout chunk (state=2): >>><<< 10710 1726776677.08660: _low_level_execute_command() done: rc=0, stdout=, stderr= 10710 1726776677.08665: _low_level_execute_command(): starting 10710 1726776677.08670: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/AnsiballZ_stat.py && sleep 0' 10710 1726776677.23280: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/profiles", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 10710 1726776677.24279: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10710 1726776677.24327: stderr chunk (state=3): >>><<< 10710 1726776677.24336: stdout chunk (state=3): >>><<< 10710 1726776677.24354: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": false}, "invocation": {"module_args": {"path": "/etc/tuned/profiles", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 10710 1726776677.24375: done with _execute_module (stat, {'path': '/etc/tuned/profiles', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10710 1726776677.24385: _low_level_execute_command(): starting 10710 1726776677.24391: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776677.0014873-10710-274625893143640/ > /dev/null 2>&1 && sleep 0' 10710 1726776677.26778: stderr chunk (state=2): >>><<< 10710 1726776677.26786: stdout chunk (state=2): >>><<< 10710 1726776677.26799: _low_level_execute_command() done: rc=0, stdout=, stderr= 10710 1726776677.26806: handler run complete 10710 1726776677.26821: attempt loop complete, returning result 10710 1726776677.26839: variable 'item' from source: unknown 10710 1726776677.26902: variable 'item' from source: unknown ok: [managed_node3] => (item=/etc/tuned/profiles) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/tuned/profiles", "stat": { "exists": false } } 10710 1726776677.26990: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776677.27000: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776677.27011: variable 'omit' from source: magic vars 10710 1726776677.27119: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10710 1726776677.27144: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10710 1726776677.27165: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10710 1726776677.27191: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10710 1726776677.27203: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10710 1726776677.27266: variable 'item' from source: unknown 10710 1726776677.27275: Evaluated conditional (item | length > 0): True 10710 1726776677.27281: variable 'omit' from source: magic vars 10710 1726776677.27292: variable 'omit' from source: magic vars 10710 1726776677.27323: variable 'item' from source: unknown 10710 1726776677.27370: variable 'item' from source: unknown 10710 1726776677.27384: variable 'omit' from source: magic vars 10710 1726776677.27400: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10710 1726776677.27409: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10710 1726776677.27416: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10710 1726776677.27428: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10710 1726776677.27434: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776677.27438: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776677.27487: Set connection var ansible_module_compression to ZIP_DEFLATED 10710 1726776677.27494: Set connection var ansible_shell_type to sh 10710 1726776677.27502: Set connection var ansible_timeout to 10 10710 1726776677.27507: Set connection var ansible_pipelining to False 10710 1726776677.27514: Set connection var ansible_shell_executable to /bin/sh 10710 1726776677.27516: Set connection var ansible_connection to ssh 10710 1726776677.27531: variable 'ansible_shell_executable' from source: unknown 10710 1726776677.27534: variable 'ansible_connection' from source: unknown 10710 1726776677.27538: variable 'ansible_module_compression' from source: unknown 10710 1726776677.27541: variable 'ansible_shell_type' from source: unknown 10710 1726776677.27544: variable 'ansible_shell_executable' from source: unknown 10710 1726776677.27547: variable 'ansible_host' from source: host vars for 'managed_node3' 10710 1726776677.27552: variable 'ansible_pipelining' from source: unknown 10710 1726776677.27555: variable 'ansible_timeout' from source: unknown 10710 1726776677.27560: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10710 1726776677.27623: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10710 1726776677.27634: variable 'omit' from source: magic vars 10710 1726776677.27640: starting attempt loop 10710 1726776677.27643: running the handler 10710 1726776677.27650: _low_level_execute_command(): starting 10710 1726776677.27655: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10710 1726776677.29805: stdout chunk (state=2): >>>/root <<< 10710 1726776677.29924: stderr chunk (state=3): >>><<< 10710 1726776677.29932: stdout chunk (state=3): >>><<< 10710 1726776677.29942: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10710 1726776677.29949: _low_level_execute_command(): starting 10710 1726776677.29954: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475 `" && echo ansible-tmp-1726776677.2994661-10710-22131299648475="` echo /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475 `" ) && sleep 0' 10710 1726776677.32347: stdout chunk (state=2): >>>ansible-tmp-1726776677.2994661-10710-22131299648475=/root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475 <<< 10710 1726776677.32790: stderr chunk (state=3): >>><<< 10710 1726776677.32797: stdout chunk (state=3): >>><<< 10710 1726776677.32814: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776677.2994661-10710-22131299648475=/root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475 , stderr= 10710 1726776677.32855: variable 'ansible_module_compression' from source: unknown 10710 1726776677.32904: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10710 1726776677.32924: variable 'ansible_facts' from source: unknown 10710 1726776677.33001: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/AnsiballZ_stat.py 10710 1726776677.33450: Sending initial data 10710 1726776677.33459: Sent initial data (151 bytes) 10710 1726776677.36350: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmps4euey33 /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/AnsiballZ_stat.py <<< 10710 1726776677.37434: stderr chunk (state=3): >>><<< 10710 1726776677.37441: stdout chunk (state=3): >>><<< 10710 1726776677.37461: done transferring module to remote 10710 1726776677.37471: _low_level_execute_command(): starting 10710 1726776677.37476: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/ /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/AnsiballZ_stat.py && sleep 0' 10710 1726776677.40186: stderr chunk (state=2): >>><<< 10710 1726776677.40195: stdout chunk (state=2): >>><<< 10710 1726776677.40212: _low_level_execute_command() done: rc=0, stdout=, stderr= 10710 1726776677.40218: _low_level_execute_command(): starting 10710 1726776677.40223: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/AnsiballZ_stat.py && sleep 0' 10710 1726776677.55949: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned", "mode": "0755", "isdir": true, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 159, "inode": 917919, "dev": 51713, "nlink": 4, "atime": 1726776658.090236, "mtime": 1726776656.4062142, "ctime": 1726776656.4062142, "wusr": true, "rusr": true, "xusr": true, "wgrp": false, "rgrp": true, "xgrp": true, "woth": false, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "pw_name": "root", "gr_name": "root", "mimetype": "inode/directory", "charset": "binary", "version": "1785990601", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} <<< 10710 1726776677.57143: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10710 1726776677.57192: stderr chunk (state=3): >>><<< 10710 1726776677.57198: stdout chunk (state=3): >>><<< 10710 1726776677.57213: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned", "mode": "0755", "isdir": true, "ischr": false, "isblk": false, "isreg": false, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 159, "inode": 917919, "dev": 51713, "nlink": 4, "atime": 1726776658.090236, "mtime": 1726776656.4062142, "ctime": 1726776656.4062142, "wusr": true, "rusr": true, "xusr": true, "wgrp": false, "rgrp": true, "xgrp": true, "woth": false, "roth": true, "xoth": true, "isuid": false, "isgid": false, "blocks": 0, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": true, "pw_name": "root", "gr_name": "root", "mimetype": "inode/directory", "charset": "binary", "version": "1785990601", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned", "follow": false, "get_checksum": true, "get_mime": true, "get_attributes": true, "checksum_algorithm": "sha1"}}} , stderr=Shared connection to 10.31.8.186 closed. 10710 1726776677.57249: done with _execute_module (stat, {'path': '/etc/tuned', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10710 1726776677.57259: _low_level_execute_command(): starting 10710 1726776677.57264: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776677.2994661-10710-22131299648475/ > /dev/null 2>&1 && sleep 0' 10710 1726776677.59813: stderr chunk (state=2): >>><<< 10710 1726776677.59825: stdout chunk (state=2): >>><<< 10710 1726776677.59846: _low_level_execute_command() done: rc=0, stdout=, stderr= 10710 1726776677.59855: handler run complete 10710 1726776677.59891: attempt loop complete, returning result 10710 1726776677.59912: variable 'item' from source: unknown 10710 1726776677.59982: variable 'item' from source: unknown ok: [managed_node3] => (item=/etc/tuned) => { "ansible_loop_var": "item", "changed": false, "item": "/etc/tuned", "stat": { "atime": 1726776658.090236, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1726776656.4062142, "dev": 51713, "device_type": 0, "executable": true, "exists": true, "gid": 0, "gr_name": "root", "inode": 917919, "isblk": false, "ischr": false, "isdir": true, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/directory", "mode": "0755", "mtime": 1726776656.4062142, "nlink": 4, "path": "/etc/tuned", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 159, "uid": 0, "version": "1785990601", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": true, "xoth": true, "xusr": true } } 10710 1726776677.60026: dumping result to json 10710 1726776677.60038: done dumping result, returning 10710 1726776677.60046: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory [120fa90a-8a95-fee7-0cea-0000000006e1] 10710 1726776677.60056: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e1 10710 1726776677.60097: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e1 10710 1726776677.60101: WORKER PROCESS EXITING 9704 1726776677.60346: no more pending results, returning what we have 9704 1726776677.60349: results queue empty 9704 1726776677.60350: checking for any_errors_fatal 9704 1726776677.60354: done checking for any_errors_fatal 9704 1726776677.60355: checking for max_fail_percentage 9704 1726776677.60356: done checking for max_fail_percentage 9704 1726776677.60357: checking to see if all hosts have failed and the running result is not ok 9704 1726776677.60357: done checking to see if all hosts have failed 9704 1726776677.60358: getting the remaining hosts for this loop 9704 1726776677.60359: done getting the remaining hosts for this loop 9704 1726776677.60362: getting the next task for host managed_node3 9704 1726776677.60367: done getting next task for host managed_node3 9704 1726776677.60370: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir 9704 1726776677.60373: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776677.60381: getting variables 9704 1726776677.60383: in VariableManager get_vars() 9704 1726776677.60408: Calling all_inventory to load vars for managed_node3 9704 1726776677.60410: Calling groups_inventory to load vars for managed_node3 9704 1726776677.60412: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776677.60420: Calling all_plugins_play to load vars for managed_node3 9704 1726776677.60423: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776677.60425: Calling groups_plugins_play to load vars for managed_node3 9704 1726776677.60540: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776677.60661: done with get_vars() 9704 1726776677.60669: done getting variables 9704 1726776677.60711: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:63 Thursday 19 September 2024 16:11:17 -0400 (0:00:00.682) 0:00:30.536 **** 9704 1726776677.60736: entering _queue_task() for managed_node3/set_fact 9704 1726776677.60922: worker is 1 (out of 1 available) 9704 1726776677.60937: exiting _queue_task() for managed_node3/set_fact 9704 1726776677.60950: done queuing things up, now waiting for results queue to drain 9704 1726776677.60952: waiting for pending results... 10741 1726776677.61109: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir 10741 1726776677.61268: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e2 10741 1726776677.61288: variable 'ansible_search_path' from source: unknown 10741 1726776677.61292: variable 'ansible_search_path' from source: unknown 10741 1726776677.61324: calling self._execute() 10741 1726776677.61418: variable 'ansible_host' from source: host vars for 'managed_node3' 10741 1726776677.61428: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10741 1726776677.61439: variable 'omit' from source: magic vars 10741 1726776677.61535: variable 'omit' from source: magic vars 10741 1726776677.61598: variable 'omit' from source: magic vars 10741 1726776677.61973: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10741 1726776677.63796: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10741 1726776677.63861: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10741 1726776677.63899: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10741 1726776677.63937: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10741 1726776677.63965: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10741 1726776677.64036: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10741 1726776677.64066: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10741 1726776677.64090: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10741 1726776677.64157: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10741 1726776677.64173: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10741 1726776677.64216: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10741 1726776677.64241: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10741 1726776677.64270: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10741 1726776677.64309: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10741 1726776677.64323: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10741 1726776677.64384: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10741 1726776677.64410: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10741 1726776677.64436: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10741 1726776677.64466: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10741 1726776677.64478: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10741 1726776677.64631: variable '__kernel_settings_find_profile_dirs' from source: set_fact 10741 1726776677.64695: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10741 1726776677.64818: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10741 1726776677.64847: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10741 1726776677.64871: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10741 1726776677.64893: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10741 1726776677.64923: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py (found_in_cache=True, class_only=False) 10741 1726776677.64942: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py (found_in_cache=True, class_only=False) 10741 1726776677.64962: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py (found_in_cache=True, class_only=False) 10741 1726776677.64981: Loading TestModule 'ansible_collections.ansible.builtin.plugins.test.uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py (found_in_cache=True, class_only=False) 10741 1726776677.65018: variable 'omit' from source: magic vars 10741 1726776677.65042: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10741 1726776677.65065: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10741 1726776677.65080: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10741 1726776677.65093: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10741 1726776677.65102: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10741 1726776677.65124: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10741 1726776677.65127: variable 'ansible_host' from source: host vars for 'managed_node3' 10741 1726776677.65132: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10741 1726776677.65195: Set connection var ansible_module_compression to ZIP_DEFLATED 10741 1726776677.65201: Set connection var ansible_shell_type to sh 10741 1726776677.65207: Set connection var ansible_timeout to 10 10741 1726776677.65211: Set connection var ansible_pipelining to False 10741 1726776677.65215: Set connection var ansible_shell_executable to /bin/sh 10741 1726776677.65217: Set connection var ansible_connection to ssh 10741 1726776677.65235: variable 'ansible_shell_executable' from source: unknown 10741 1726776677.65239: variable 'ansible_connection' from source: unknown 10741 1726776677.65243: variable 'ansible_module_compression' from source: unknown 10741 1726776677.65246: variable 'ansible_shell_type' from source: unknown 10741 1726776677.65249: variable 'ansible_shell_executable' from source: unknown 10741 1726776677.65255: variable 'ansible_host' from source: host vars for 'managed_node3' 10741 1726776677.65259: variable 'ansible_pipelining' from source: unknown 10741 1726776677.65262: variable 'ansible_timeout' from source: unknown 10741 1726776677.65266: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10741 1726776677.65327: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10741 1726776677.65339: variable 'omit' from source: magic vars 10741 1726776677.65345: starting attempt loop 10741 1726776677.65349: running the handler 10741 1726776677.65361: handler run complete 10741 1726776677.65369: attempt loop complete, returning result 10741 1726776677.65372: _execute() done 10741 1726776677.65375: dumping result to json 10741 1726776677.65379: done dumping result, returning 10741 1726776677.65385: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set tuned profile parent dir [120fa90a-8a95-fee7-0cea-0000000006e2] 10741 1726776677.65390: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e2 10741 1726776677.65410: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e2 10741 1726776677.65413: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_profile_parent": "/etc/tuned" }, "changed": false } 9704 1726776677.65547: no more pending results, returning what we have 9704 1726776677.65550: results queue empty 9704 1726776677.65550: checking for any_errors_fatal 9704 1726776677.65565: done checking for any_errors_fatal 9704 1726776677.65565: checking for max_fail_percentage 9704 1726776677.65566: done checking for max_fail_percentage 9704 1726776677.65567: checking to see if all hosts have failed and the running result is not ok 9704 1726776677.65568: done checking to see if all hosts have failed 9704 1726776677.65568: getting the remaining hosts for this loop 9704 1726776677.65569: done getting the remaining hosts for this loop 9704 1726776677.65573: getting the next task for host managed_node3 9704 1726776677.65578: done getting next task for host managed_node3 9704 1726776677.65582: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started 9704 1726776677.65584: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776677.65595: getting variables 9704 1726776677.65597: in VariableManager get_vars() 9704 1726776677.65630: Calling all_inventory to load vars for managed_node3 9704 1726776677.65633: Calling groups_inventory to load vars for managed_node3 9704 1726776677.65635: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776677.65643: Calling all_plugins_play to load vars for managed_node3 9704 1726776677.65645: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776677.65647: Calling groups_plugins_play to load vars for managed_node3 9704 1726776677.65765: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776677.65889: done with get_vars() 9704 1726776677.65897: done getting variables 9704 1726776677.65940: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:67 Thursday 19 September 2024 16:11:17 -0400 (0:00:00.052) 0:00:30.588 **** 9704 1726776677.65965: entering _queue_task() for managed_node3/service 9704 1726776677.66125: worker is 1 (out of 1 available) 9704 1726776677.66142: exiting _queue_task() for managed_node3/service 9704 1726776677.66154: done queuing things up, now waiting for results queue to drain 9704 1726776677.66156: waiting for pending results... 10744 1726776677.66284: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started 10744 1726776677.66404: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e3 10744 1726776677.66420: variable 'ansible_search_path' from source: unknown 10744 1726776677.66424: variable 'ansible_search_path' from source: unknown 10744 1726776677.66460: variable '__kernel_settings_services' from source: include_vars 10744 1726776677.66761: variable '__kernel_settings_services' from source: include_vars 10744 1726776677.66815: variable 'omit' from source: magic vars 10744 1726776677.66886: variable 'ansible_host' from source: host vars for 'managed_node3' 10744 1726776677.66896: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10744 1726776677.66905: variable 'omit' from source: magic vars 10744 1726776677.66962: variable 'omit' from source: magic vars 10744 1726776677.66997: variable 'omit' from source: magic vars 10744 1726776677.67032: variable 'item' from source: unknown 10744 1726776677.67087: variable 'item' from source: unknown 10744 1726776677.67107: variable 'omit' from source: magic vars 10744 1726776677.67138: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10744 1726776677.67166: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10744 1726776677.67184: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10744 1726776677.67197: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10744 1726776677.67207: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10744 1726776677.67230: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10744 1726776677.67236: variable 'ansible_host' from source: host vars for 'managed_node3' 10744 1726776677.67240: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10744 1726776677.67304: Set connection var ansible_module_compression to ZIP_DEFLATED 10744 1726776677.67310: Set connection var ansible_shell_type to sh 10744 1726776677.67315: Set connection var ansible_timeout to 10 10744 1726776677.67318: Set connection var ansible_pipelining to False 10744 1726776677.67323: Set connection var ansible_shell_executable to /bin/sh 10744 1726776677.67324: Set connection var ansible_connection to ssh 10744 1726776677.67341: variable 'ansible_shell_executable' from source: unknown 10744 1726776677.67346: variable 'ansible_connection' from source: unknown 10744 1726776677.67350: variable 'ansible_module_compression' from source: unknown 10744 1726776677.67356: variable 'ansible_shell_type' from source: unknown 10744 1726776677.67360: variable 'ansible_shell_executable' from source: unknown 10744 1726776677.67364: variable 'ansible_host' from source: host vars for 'managed_node3' 10744 1726776677.67367: variable 'ansible_pipelining' from source: unknown 10744 1726776677.67371: variable 'ansible_timeout' from source: unknown 10744 1726776677.67375: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10744 1726776677.67465: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10744 1726776677.67477: variable 'omit' from source: magic vars 10744 1726776677.67483: starting attempt loop 10744 1726776677.67487: running the handler 10744 1726776677.67546: variable 'ansible_facts' from source: unknown 10744 1726776677.67623: _low_level_execute_command(): starting 10744 1726776677.67634: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10744 1726776677.69966: stdout chunk (state=2): >>>/root <<< 10744 1726776677.70095: stderr chunk (state=3): >>><<< 10744 1726776677.70103: stdout chunk (state=3): >>><<< 10744 1726776677.70124: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10744 1726776677.70139: _low_level_execute_command(): starting 10744 1726776677.70147: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567 `" && echo ansible-tmp-1726776677.7013304-10744-254244533859567="` echo /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567 `" ) && sleep 0' 10744 1726776677.72665: stdout chunk (state=2): >>>ansible-tmp-1726776677.7013304-10744-254244533859567=/root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567 <<< 10744 1726776677.72798: stderr chunk (state=3): >>><<< 10744 1726776677.72805: stdout chunk (state=3): >>><<< 10744 1726776677.72819: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776677.7013304-10744-254244533859567=/root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567 , stderr= 10744 1726776677.72846: variable 'ansible_module_compression' from source: unknown 10744 1726776677.72890: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 10744 1726776677.72937: variable 'ansible_facts' from source: unknown 10744 1726776677.73096: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/AnsiballZ_systemd.py 10744 1726776677.73206: Sending initial data 10744 1726776677.73214: Sent initial data (155 bytes) 10744 1726776677.75745: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmplzpbvtz2 /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/AnsiballZ_systemd.py <<< 10744 1726776677.77591: stderr chunk (state=3): >>><<< 10744 1726776677.77598: stdout chunk (state=3): >>><<< 10744 1726776677.77618: done transferring module to remote 10744 1726776677.77631: _low_level_execute_command(): starting 10744 1726776677.77637: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/ /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/AnsiballZ_systemd.py && sleep 0' 10744 1726776677.80038: stderr chunk (state=2): >>><<< 10744 1726776677.80045: stdout chunk (state=2): >>><<< 10744 1726776677.80061: _low_level_execute_command() done: rc=0, stdout=, stderr= 10744 1726776677.80066: _low_level_execute_command(): starting 10744 1726776677.80071: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/AnsiballZ_systemd.py && sleep 0' 10744 1726776678.08286: stdout chunk (state=2): >>> {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "15002", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15069184", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "Memo<<< 10744 1726776678.08350: stdout chunk (state=3): >>>ryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 10744 1726776678.10340: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10744 1726776678.10385: stderr chunk (state=3): >>><<< 10744 1726776678.10392: stdout chunk (state=3): >>><<< 10744 1726776678.10416: _low_level_execute_command() done: rc=0, stdout= {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "15002", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15069184", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10744 1726776678.10903: done with _execute_module (ansible.legacy.systemd, {'name': 'tuned', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10744 1726776678.10924: _low_level_execute_command(): starting 10744 1726776678.10932: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776677.7013304-10744-254244533859567/ > /dev/null 2>&1 && sleep 0' 10744 1726776678.14495: stderr chunk (state=2): >>><<< 10744 1726776678.14506: stdout chunk (state=2): >>><<< 10744 1726776678.14521: _low_level_execute_command() done: rc=0, stdout=, stderr= 10744 1726776678.14527: handler run complete 10744 1726776678.14579: attempt loop complete, returning result 10744 1726776678.14597: variable 'item' from source: unknown 10744 1726776678.14675: variable 'item' from source: unknown ok: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": false, "enabled": true, "item": "tuned", "name": "tuned", "state": "started", "status": { "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "ActiveState": "active", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "com.redhat.tuned", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "ConfigurationDirectoryMode": "0755", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "ControlGroup": "/system.slice/tuned.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Dynamic System Tuning Daemon", "DevicePolicy": "auto", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "tuned.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "15002", "MemoryAccounting": "yes", "MemoryCurrent": "15069184", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "tuned.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/tuned/tuned.pid", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "WatchdogUSec": "0" } } 10744 1726776678.14807: dumping result to json 10744 1726776678.14824: done dumping result, returning 10744 1726776678.14833: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started [120fa90a-8a95-fee7-0cea-0000000006e3] 10744 1726776678.14838: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e3 10744 1726776678.14942: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e3 10744 1726776678.14947: WORKER PROCESS EXITING 9704 1726776678.16349: no more pending results, returning what we have 9704 1726776678.16352: results queue empty 9704 1726776678.16356: checking for any_errors_fatal 9704 1726776678.16359: done checking for any_errors_fatal 9704 1726776678.16360: checking for max_fail_percentage 9704 1726776678.16361: done checking for max_fail_percentage 9704 1726776678.16362: checking to see if all hosts have failed and the running result is not ok 9704 1726776678.16362: done checking to see if all hosts have failed 9704 1726776678.16363: getting the remaining hosts for this loop 9704 1726776678.16364: done getting the remaining hosts for this loop 9704 1726776678.16367: getting the next task for host managed_node3 9704 1726776678.16373: done getting next task for host managed_node3 9704 1726776678.16376: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists 9704 1726776678.16379: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776678.16389: getting variables 9704 1726776678.16391: in VariableManager get_vars() 9704 1726776678.16418: Calling all_inventory to load vars for managed_node3 9704 1726776678.16421: Calling groups_inventory to load vars for managed_node3 9704 1726776678.16423: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776678.16434: Calling all_plugins_play to load vars for managed_node3 9704 1726776678.16437: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776678.16440: Calling groups_plugins_play to load vars for managed_node3 9704 1726776678.16596: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776678.16795: done with get_vars() 9704 1726776678.16805: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:74 Thursday 19 September 2024 16:11:18 -0400 (0:00:00.509) 0:00:31.097 **** 9704 1726776678.16903: entering _queue_task() for managed_node3/file 9704 1726776678.17104: worker is 1 (out of 1 available) 9704 1726776678.17118: exiting _queue_task() for managed_node3/file 9704 1726776678.17132: done queuing things up, now waiting for results queue to drain 9704 1726776678.17135: waiting for pending results... 10767 1726776678.17269: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists 10767 1726776678.17408: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e4 10767 1726776678.17425: variable 'ansible_search_path' from source: unknown 10767 1726776678.17431: variable 'ansible_search_path' from source: unknown 10767 1726776678.17457: calling self._execute() 10767 1726776678.17525: variable 'ansible_host' from source: host vars for 'managed_node3' 10767 1726776678.17535: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10767 1726776678.17542: variable 'omit' from source: magic vars 10767 1726776678.17612: variable 'omit' from source: magic vars 10767 1726776678.17658: variable 'omit' from source: magic vars 10767 1726776678.17680: variable '__kernel_settings_profile_dir' from source: role '' all vars 10767 1726776678.17899: variable '__kernel_settings_profile_dir' from source: role '' all vars 10767 1726776678.17974: variable '__kernel_settings_profile_parent' from source: set_fact 10767 1726776678.17983: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10767 1726776678.18042: variable 'omit' from source: magic vars 10767 1726776678.18073: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10767 1726776678.18100: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10767 1726776678.18118: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10767 1726776678.18135: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10767 1726776678.18145: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10767 1726776678.18168: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10767 1726776678.18173: variable 'ansible_host' from source: host vars for 'managed_node3' 10767 1726776678.18177: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10767 1726776678.18245: Set connection var ansible_module_compression to ZIP_DEFLATED 10767 1726776678.18253: Set connection var ansible_shell_type to sh 10767 1726776678.18261: Set connection var ansible_timeout to 10 10767 1726776678.18267: Set connection var ansible_pipelining to False 10767 1726776678.18274: Set connection var ansible_shell_executable to /bin/sh 10767 1726776678.18277: Set connection var ansible_connection to ssh 10767 1726776678.18291: variable 'ansible_shell_executable' from source: unknown 10767 1726776678.18295: variable 'ansible_connection' from source: unknown 10767 1726776678.18298: variable 'ansible_module_compression' from source: unknown 10767 1726776678.18301: variable 'ansible_shell_type' from source: unknown 10767 1726776678.18304: variable 'ansible_shell_executable' from source: unknown 10767 1726776678.18311: variable 'ansible_host' from source: host vars for 'managed_node3' 10767 1726776678.18316: variable 'ansible_pipelining' from source: unknown 10767 1726776678.18319: variable 'ansible_timeout' from source: unknown 10767 1726776678.18323: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10767 1726776678.18457: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10767 1726776678.18468: variable 'omit' from source: magic vars 10767 1726776678.18474: starting attempt loop 10767 1726776678.18478: running the handler 10767 1726776678.18489: _low_level_execute_command(): starting 10767 1726776678.18496: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10767 1726776678.20805: stdout chunk (state=2): >>>/root <<< 10767 1726776678.20919: stderr chunk (state=3): >>><<< 10767 1726776678.20926: stdout chunk (state=3): >>><<< 10767 1726776678.20945: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10767 1726776678.20959: _low_level_execute_command(): starting 10767 1726776678.20965: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325 `" && echo ansible-tmp-1726776678.2095213-10767-94059150329325="` echo /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325 `" ) && sleep 0' 10767 1726776678.23367: stdout chunk (state=2): >>>ansible-tmp-1726776678.2095213-10767-94059150329325=/root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325 <<< 10767 1726776678.23492: stderr chunk (state=3): >>><<< 10767 1726776678.23498: stdout chunk (state=3): >>><<< 10767 1726776678.23514: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776678.2095213-10767-94059150329325=/root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325 , stderr= 10767 1726776678.23549: variable 'ansible_module_compression' from source: unknown 10767 1726776678.23592: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10767 1726776678.23625: variable 'ansible_facts' from source: unknown 10767 1726776678.23696: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/AnsiballZ_file.py 10767 1726776678.23791: Sending initial data 10767 1726776678.23798: Sent initial data (151 bytes) 10767 1726776678.26193: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpxdseeluq /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/AnsiballZ_file.py <<< 10767 1726776678.27378: stderr chunk (state=3): >>><<< 10767 1726776678.27383: stdout chunk (state=3): >>><<< 10767 1726776678.27396: done transferring module to remote 10767 1726776678.27404: _low_level_execute_command(): starting 10767 1726776678.27407: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/ /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/AnsiballZ_file.py && sleep 0' 10767 1726776678.29713: stderr chunk (state=2): >>><<< 10767 1726776678.29720: stdout chunk (state=2): >>><<< 10767 1726776678.29733: _low_level_execute_command() done: rc=0, stdout=, stderr= 10767 1726776678.29738: _low_level_execute_command(): starting 10767 1726776678.29743: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/AnsiballZ_file.py && sleep 0' 10767 1726776678.45695: stdout chunk (state=2): >>> {"path": "/etc/tuned/kernel_settings", "changed": false, "diff": {"before": {"path": "/etc/tuned/kernel_settings"}, "after": {"path": "/etc/tuned/kernel_settings"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0755", "state": "directory", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 24, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "directory", "mode": "0755", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10767 1726776678.46844: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10767 1726776678.46894: stderr chunk (state=3): >>><<< 10767 1726776678.46909: stdout chunk (state=3): >>><<< 10767 1726776678.46923: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/kernel_settings", "changed": false, "diff": {"before": {"path": "/etc/tuned/kernel_settings"}, "after": {"path": "/etc/tuned/kernel_settings"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0755", "state": "directory", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 24, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "directory", "mode": "0755", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10767 1726776678.46959: done with _execute_module (file, {'path': '/etc/tuned/kernel_settings', 'state': 'directory', 'mode': '0755', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10767 1726776678.46970: _low_level_execute_command(): starting 10767 1726776678.46976: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776678.2095213-10767-94059150329325/ > /dev/null 2>&1 && sleep 0' 10767 1726776678.49377: stderr chunk (state=2): >>><<< 10767 1726776678.49384: stdout chunk (state=2): >>><<< 10767 1726776678.49398: _low_level_execute_command() done: rc=0, stdout=, stderr= 10767 1726776678.49405: handler run complete 10767 1726776678.49423: attempt loop complete, returning result 10767 1726776678.49427: _execute() done 10767 1726776678.49432: dumping result to json 10767 1726776678.49438: done dumping result, returning 10767 1726776678.49445: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel settings profile directory exists [120fa90a-8a95-fee7-0cea-0000000006e4] 10767 1726776678.49451: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e4 10767 1726776678.49489: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e4 10767 1726776678.49493: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "gid": 0, "group": "root", "mode": "0755", "owner": "root", "path": "/etc/tuned/kernel_settings", "secontext": "unconfined_u:object_r:tuned_etc_t:s0", "size": 24, "state": "directory", "uid": 0 } 9704 1726776678.49647: no more pending results, returning what we have 9704 1726776678.49650: results queue empty 9704 1726776678.49651: checking for any_errors_fatal 9704 1726776678.49669: done checking for any_errors_fatal 9704 1726776678.49670: checking for max_fail_percentage 9704 1726776678.49671: done checking for max_fail_percentage 9704 1726776678.49672: checking to see if all hosts have failed and the running result is not ok 9704 1726776678.49673: done checking to see if all hosts have failed 9704 1726776678.49673: getting the remaining hosts for this loop 9704 1726776678.49674: done getting the remaining hosts for this loop 9704 1726776678.49678: getting the next task for host managed_node3 9704 1726776678.49683: done getting next task for host managed_node3 9704 1726776678.49686: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get active_profile 9704 1726776678.49690: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776678.49699: getting variables 9704 1726776678.49701: in VariableManager get_vars() 9704 1726776678.49735: Calling all_inventory to load vars for managed_node3 9704 1726776678.49738: Calling groups_inventory to load vars for managed_node3 9704 1726776678.49739: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776678.49748: Calling all_plugins_play to load vars for managed_node3 9704 1726776678.49750: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776678.49752: Calling groups_plugins_play to load vars for managed_node3 9704 1726776678.49871: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776678.50006: done with get_vars() 9704 1726776678.50014: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Get active_profile] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:80 Thursday 19 September 2024 16:11:18 -0400 (0:00:00.331) 0:00:31.429 **** 9704 1726776678.50087: entering _queue_task() for managed_node3/slurp 9704 1726776678.50256: worker is 1 (out of 1 available) 9704 1726776678.50271: exiting _queue_task() for managed_node3/slurp 9704 1726776678.50282: done queuing things up, now waiting for results queue to drain 9704 1726776678.50284: waiting for pending results... 10780 1726776678.50409: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get active_profile 10780 1726776678.50534: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e5 10780 1726776678.50551: variable 'ansible_search_path' from source: unknown 10780 1726776678.50555: variable 'ansible_search_path' from source: unknown 10780 1726776678.50581: calling self._execute() 10780 1726776678.50654: variable 'ansible_host' from source: host vars for 'managed_node3' 10780 1726776678.50661: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10780 1726776678.50667: variable 'omit' from source: magic vars 10780 1726776678.50737: variable 'omit' from source: magic vars 10780 1726776678.50773: variable 'omit' from source: magic vars 10780 1726776678.50793: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10780 1726776678.51015: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10780 1726776678.51077: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10780 1726776678.51105: variable 'omit' from source: magic vars 10780 1726776678.51136: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10780 1726776678.51227: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10780 1726776678.51249: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10780 1726776678.51263: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10780 1726776678.51273: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10780 1726776678.51299: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10780 1726776678.51305: variable 'ansible_host' from source: host vars for 'managed_node3' 10780 1726776678.51309: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10780 1726776678.51376: Set connection var ansible_module_compression to ZIP_DEFLATED 10780 1726776678.51384: Set connection var ansible_shell_type to sh 10780 1726776678.51392: Set connection var ansible_timeout to 10 10780 1726776678.51398: Set connection var ansible_pipelining to False 10780 1726776678.51403: Set connection var ansible_shell_executable to /bin/sh 10780 1726776678.51405: Set connection var ansible_connection to ssh 10780 1726776678.51419: variable 'ansible_shell_executable' from source: unknown 10780 1726776678.51422: variable 'ansible_connection' from source: unknown 10780 1726776678.51423: variable 'ansible_module_compression' from source: unknown 10780 1726776678.51425: variable 'ansible_shell_type' from source: unknown 10780 1726776678.51427: variable 'ansible_shell_executable' from source: unknown 10780 1726776678.51442: variable 'ansible_host' from source: host vars for 'managed_node3' 10780 1726776678.51446: variable 'ansible_pipelining' from source: unknown 10780 1726776678.51450: variable 'ansible_timeout' from source: unknown 10780 1726776678.51454: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10780 1726776678.51584: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10780 1726776678.51594: variable 'omit' from source: magic vars 10780 1726776678.51600: starting attempt loop 10780 1726776678.51603: running the handler 10780 1726776678.51616: _low_level_execute_command(): starting 10780 1726776678.51624: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10780 1726776678.53909: stdout chunk (state=2): >>>/root <<< 10780 1726776678.54026: stderr chunk (state=3): >>><<< 10780 1726776678.54034: stdout chunk (state=3): >>><<< 10780 1726776678.54050: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10780 1726776678.54065: _low_level_execute_command(): starting 10780 1726776678.54071: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404 `" && echo ansible-tmp-1726776678.5405982-10780-192889062348404="` echo /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404 `" ) && sleep 0' 10780 1726776678.56493: stdout chunk (state=2): >>>ansible-tmp-1726776678.5405982-10780-192889062348404=/root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404 <<< 10780 1726776678.56618: stderr chunk (state=3): >>><<< 10780 1726776678.56625: stdout chunk (state=3): >>><<< 10780 1726776678.56641: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776678.5405982-10780-192889062348404=/root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404 , stderr= 10780 1726776678.56680: variable 'ansible_module_compression' from source: unknown 10780 1726776678.56713: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.slurp-ZIP_DEFLATED 10780 1726776678.56742: variable 'ansible_facts' from source: unknown 10780 1726776678.56814: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/AnsiballZ_slurp.py 10780 1726776678.56913: Sending initial data 10780 1726776678.56921: Sent initial data (153 bytes) 10780 1726776678.59377: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp6uvuq3am /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/AnsiballZ_slurp.py <<< 10780 1726776678.60336: stderr chunk (state=3): >>><<< 10780 1726776678.60346: stdout chunk (state=3): >>><<< 10780 1726776678.60367: done transferring module to remote 10780 1726776678.60379: _low_level_execute_command(): starting 10780 1726776678.60384: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/ /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/AnsiballZ_slurp.py && sleep 0' 10780 1726776678.62758: stderr chunk (state=2): >>><<< 10780 1726776678.62767: stdout chunk (state=2): >>><<< 10780 1726776678.62784: _low_level_execute_command() done: rc=0, stdout=, stderr= 10780 1726776678.62790: _low_level_execute_command(): starting 10780 1726776678.62796: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/AnsiballZ_slurp.py && sleep 0' 10780 1726776678.77844: stdout chunk (state=2): >>> {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} <<< 10780 1726776678.78816: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10780 1726776678.78866: stderr chunk (state=3): >>><<< 10780 1726776678.78873: stdout chunk (state=3): >>><<< 10780 1726776678.78889: _low_level_execute_command() done: rc=0, stdout= {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} , stderr=Shared connection to 10.31.8.186 closed. 10780 1726776678.78913: done with _execute_module (slurp, {'path': '/etc/tuned/active_profile', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10780 1726776678.78924: _low_level_execute_command(): starting 10780 1726776678.78932: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776678.5405982-10780-192889062348404/ > /dev/null 2>&1 && sleep 0' 10780 1726776678.81726: stderr chunk (state=2): >>><<< 10780 1726776678.81734: stdout chunk (state=2): >>><<< 10780 1726776678.81747: _low_level_execute_command() done: rc=0, stdout=, stderr= 10780 1726776678.81754: handler run complete 10780 1726776678.81767: attempt loop complete, returning result 10780 1726776678.81771: _execute() done 10780 1726776678.81774: dumping result to json 10780 1726776678.81778: done dumping result, returning 10780 1726776678.81785: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get active_profile [120fa90a-8a95-fee7-0cea-0000000006e5] 10780 1726776678.81791: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e5 10780 1726776678.81820: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e5 10780 1726776678.81824: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "encoding": "base64", "source": "/etc/tuned/active_profile" } 9704 1726776678.81992: no more pending results, returning what we have 9704 1726776678.81995: results queue empty 9704 1726776678.81996: checking for any_errors_fatal 9704 1726776678.82005: done checking for any_errors_fatal 9704 1726776678.82005: checking for max_fail_percentage 9704 1726776678.82006: done checking for max_fail_percentage 9704 1726776678.82007: checking to see if all hosts have failed and the running result is not ok 9704 1726776678.82008: done checking to see if all hosts have failed 9704 1726776678.82008: getting the remaining hosts for this loop 9704 1726776678.82009: done getting the remaining hosts for this loop 9704 1726776678.82012: getting the next task for host managed_node3 9704 1726776678.82018: done getting next task for host managed_node3 9704 1726776678.82022: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set active_profile 9704 1726776678.82025: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=7, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776678.82036: getting variables 9704 1726776678.82037: in VariableManager get_vars() 9704 1726776678.82072: Calling all_inventory to load vars for managed_node3 9704 1726776678.82075: Calling groups_inventory to load vars for managed_node3 9704 1726776678.82076: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776678.82083: Calling all_plugins_play to load vars for managed_node3 9704 1726776678.82085: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776678.82086: Calling groups_plugins_play to load vars for managed_node3 9704 1726776678.82201: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776678.82377: done with get_vars() 9704 1726776678.82385: done getting variables 9704 1726776678.82427: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set active_profile] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:85 Thursday 19 September 2024 16:11:18 -0400 (0:00:00.323) 0:00:31.753 **** 9704 1726776678.82452: entering _queue_task() for managed_node3/set_fact 9704 1726776678.82611: worker is 1 (out of 1 available) 9704 1726776678.82626: exiting _queue_task() for managed_node3/set_fact 9704 1726776678.82640: done queuing things up, now waiting for results queue to drain 9704 1726776678.82642: waiting for pending results... 10791 1726776678.82767: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set active_profile 10791 1726776678.82887: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e6 10791 1726776678.82904: variable 'ansible_search_path' from source: unknown 10791 1726776678.82908: variable 'ansible_search_path' from source: unknown 10791 1726776678.82937: calling self._execute() 10791 1726776678.83008: variable 'ansible_host' from source: host vars for 'managed_node3' 10791 1726776678.83017: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10791 1726776678.83027: variable 'omit' from source: magic vars 10791 1726776678.83102: variable 'omit' from source: magic vars 10791 1726776678.83140: variable 'omit' from source: magic vars 10791 1726776678.83436: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10791 1726776678.83446: variable '__cur_profile' from source: task vars 10791 1726776678.83549: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10791 1726776678.85033: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10791 1726776678.85086: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10791 1726776678.85109: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10791 1726776678.85136: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10791 1726776678.85160: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10791 1726776678.85210: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10791 1726776678.85234: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10791 1726776678.85254: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10791 1726776678.85286: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10791 1726776678.85299: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10791 1726776678.85374: variable '__kernel_settings_tuned_current_profile' from source: set_fact 10791 1726776678.85412: variable 'omit' from source: magic vars 10791 1726776678.85432: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10791 1726776678.85452: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10791 1726776678.85466: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10791 1726776678.85478: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10791 1726776678.85486: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10791 1726776678.85507: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10791 1726776678.85510: variable 'ansible_host' from source: host vars for 'managed_node3' 10791 1726776678.85512: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10791 1726776678.85572: Set connection var ansible_module_compression to ZIP_DEFLATED 10791 1726776678.85579: Set connection var ansible_shell_type to sh 10791 1726776678.85584: Set connection var ansible_timeout to 10 10791 1726776678.85587: Set connection var ansible_pipelining to False 10791 1726776678.85593: Set connection var ansible_shell_executable to /bin/sh 10791 1726776678.85595: Set connection var ansible_connection to ssh 10791 1726776678.85608: variable 'ansible_shell_executable' from source: unknown 10791 1726776678.85610: variable 'ansible_connection' from source: unknown 10791 1726776678.85612: variable 'ansible_module_compression' from source: unknown 10791 1726776678.85614: variable 'ansible_shell_type' from source: unknown 10791 1726776678.85616: variable 'ansible_shell_executable' from source: unknown 10791 1726776678.85617: variable 'ansible_host' from source: host vars for 'managed_node3' 10791 1726776678.85619: variable 'ansible_pipelining' from source: unknown 10791 1726776678.85621: variable 'ansible_timeout' from source: unknown 10791 1726776678.85623: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10791 1726776678.85695: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10791 1726776678.85705: variable 'omit' from source: magic vars 10791 1726776678.85709: starting attempt loop 10791 1726776678.85712: running the handler 10791 1726776678.85719: handler run complete 10791 1726776678.85724: attempt loop complete, returning result 10791 1726776678.85726: _execute() done 10791 1726776678.85728: dumping result to json 10791 1726776678.85732: done dumping result, returning 10791 1726776678.85736: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set active_profile [120fa90a-8a95-fee7-0cea-0000000006e6] 10791 1726776678.85740: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e6 10791 1726776678.85755: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e6 10791 1726776678.85758: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_active_profile": "virtual-guest kernel_settings" }, "changed": false } 9704 1726776678.85963: no more pending results, returning what we have 9704 1726776678.85966: results queue empty 9704 1726776678.85966: checking for any_errors_fatal 9704 1726776678.85971: done checking for any_errors_fatal 9704 1726776678.85971: checking for max_fail_percentage 9704 1726776678.85972: done checking for max_fail_percentage 9704 1726776678.85973: checking to see if all hosts have failed and the running result is not ok 9704 1726776678.85973: done checking to see if all hosts have failed 9704 1726776678.85973: getting the remaining hosts for this loop 9704 1726776678.85974: done getting the remaining hosts for this loop 9704 1726776678.85976: getting the next task for host managed_node3 9704 1726776678.85981: done getting next task for host managed_node3 9704 1726776678.85984: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile 9704 1726776678.85986: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=8, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776678.85997: getting variables 9704 1726776678.85998: in VariableManager get_vars() 9704 1726776678.86024: Calling all_inventory to load vars for managed_node3 9704 1726776678.86026: Calling groups_inventory to load vars for managed_node3 9704 1726776678.86027: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776678.86035: Calling all_plugins_play to load vars for managed_node3 9704 1726776678.86037: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776678.86039: Calling groups_plugins_play to load vars for managed_node3 9704 1726776678.86145: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776678.86275: done with get_vars() 9704 1726776678.86283: done getting variables 9704 1726776678.86320: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:91 Thursday 19 September 2024 16:11:18 -0400 (0:00:00.038) 0:00:31.792 **** 9704 1726776678.86344: entering _queue_task() for managed_node3/copy 9704 1726776678.86500: worker is 1 (out of 1 available) 9704 1726776678.86514: exiting _queue_task() for managed_node3/copy 9704 1726776678.86525: done queuing things up, now waiting for results queue to drain 9704 1726776678.86528: waiting for pending results... 10792 1726776678.86651: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile 10792 1726776678.86762: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e7 10792 1726776678.86779: variable 'ansible_search_path' from source: unknown 10792 1726776678.86782: variable 'ansible_search_path' from source: unknown 10792 1726776678.86810: calling self._execute() 10792 1726776678.86882: variable 'ansible_host' from source: host vars for 'managed_node3' 10792 1726776678.86891: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10792 1726776678.86900: variable 'omit' from source: magic vars 10792 1726776678.86975: variable 'omit' from source: magic vars 10792 1726776678.87012: variable 'omit' from source: magic vars 10792 1726776678.87036: variable '__kernel_settings_active_profile' from source: set_fact 10792 1726776678.87249: variable '__kernel_settings_active_profile' from source: set_fact 10792 1726776678.87270: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10792 1726776678.87322: variable '__kernel_settings_tuned_active_profile' from source: role '' all vars 10792 1726776678.87376: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10792 1726776678.87461: variable 'omit' from source: magic vars 10792 1726776678.87492: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10792 1726776678.87521: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10792 1726776678.87541: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10792 1726776678.87555: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10792 1726776678.87565: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10792 1726776678.87587: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10792 1726776678.87591: variable 'ansible_host' from source: host vars for 'managed_node3' 10792 1726776678.87596: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10792 1726776678.87664: Set connection var ansible_module_compression to ZIP_DEFLATED 10792 1726776678.87672: Set connection var ansible_shell_type to sh 10792 1726776678.87680: Set connection var ansible_timeout to 10 10792 1726776678.87685: Set connection var ansible_pipelining to False 10792 1726776678.87692: Set connection var ansible_shell_executable to /bin/sh 10792 1726776678.87695: Set connection var ansible_connection to ssh 10792 1726776678.87710: variable 'ansible_shell_executable' from source: unknown 10792 1726776678.87714: variable 'ansible_connection' from source: unknown 10792 1726776678.87717: variable 'ansible_module_compression' from source: unknown 10792 1726776678.87720: variable 'ansible_shell_type' from source: unknown 10792 1726776678.87724: variable 'ansible_shell_executable' from source: unknown 10792 1726776678.87726: variable 'ansible_host' from source: host vars for 'managed_node3' 10792 1726776678.87738: variable 'ansible_pipelining' from source: unknown 10792 1726776678.87742: variable 'ansible_timeout' from source: unknown 10792 1726776678.87746: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10792 1726776678.87831: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10792 1726776678.87843: variable 'omit' from source: magic vars 10792 1726776678.87850: starting attempt loop 10792 1726776678.87854: running the handler 10792 1726776678.87865: _low_level_execute_command(): starting 10792 1726776678.87872: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10792 1726776678.90193: stdout chunk (state=2): >>>/root <<< 10792 1726776678.90319: stderr chunk (state=3): >>><<< 10792 1726776678.90327: stdout chunk (state=3): >>><<< 10792 1726776678.90347: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10792 1726776678.90361: _low_level_execute_command(): starting 10792 1726776678.90368: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387 `" && echo ansible-tmp-1726776678.9035478-10792-266422440600387="` echo /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387 `" ) && sleep 0' 10792 1726776678.92795: stdout chunk (state=2): >>>ansible-tmp-1726776678.9035478-10792-266422440600387=/root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387 <<< 10792 1726776678.92917: stderr chunk (state=3): >>><<< 10792 1726776678.92924: stdout chunk (state=3): >>><<< 10792 1726776678.92938: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776678.9035478-10792-266422440600387=/root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387 , stderr= 10792 1726776678.93008: variable 'ansible_module_compression' from source: unknown 10792 1726776678.93052: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10792 1726776678.93083: variable 'ansible_facts' from source: unknown 10792 1726776678.93151: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_stat.py 10792 1726776678.93236: Sending initial data 10792 1726776678.93244: Sent initial data (152 bytes) 10792 1726776678.95641: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpkkoii1_q /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_stat.py <<< 10792 1726776678.96603: stderr chunk (state=3): >>><<< 10792 1726776678.96609: stdout chunk (state=3): >>><<< 10792 1726776678.96625: done transferring module to remote 10792 1726776678.96635: _low_level_execute_command(): starting 10792 1726776678.96640: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/ /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_stat.py && sleep 0' 10792 1726776678.98936: stderr chunk (state=2): >>><<< 10792 1726776678.98942: stdout chunk (state=2): >>><<< 10792 1726776678.98954: _low_level_execute_command() done: rc=0, stdout=, stderr= 10792 1726776678.98960: _low_level_execute_command(): starting 10792 1726776678.98964: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_stat.py && sleep 0' 10792 1726776679.15135: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 30, "inode": 142606531, "dev": 51713, "nlink": 1, "atime": 1726776663.9393117, "mtime": 1726776658.1132364, "ctime": 1726776658.1132364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "mimetype": "text/plain", "charset": "us-ascii", "version": "3590528432", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10792 1726776679.16322: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10792 1726776679.16372: stderr chunk (state=3): >>><<< 10792 1726776679.16380: stdout chunk (state=3): >>><<< 10792 1726776679.16397: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 30, "inode": 142606531, "dev": 51713, "nlink": 1, "atime": 1726776663.9393117, "mtime": 1726776658.1132364, "ctime": 1726776658.1132364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "mimetype": "text/plain", "charset": "us-ascii", "version": "3590528432", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10792 1726776679.16441: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/active_profile', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10792 1726776679.16480: variable 'ansible_module_compression' from source: unknown 10792 1726776679.16513: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10792 1726776679.16535: variable 'ansible_facts' from source: unknown 10792 1726776679.16594: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_file.py 10792 1726776679.16688: Sending initial data 10792 1726776679.16695: Sent initial data (152 bytes) 10792 1726776679.19226: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpp3we3dze /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_file.py <<< 10792 1726776679.20244: stderr chunk (state=3): >>><<< 10792 1726776679.20251: stdout chunk (state=3): >>><<< 10792 1726776679.20268: done transferring module to remote 10792 1726776679.20277: _low_level_execute_command(): starting 10792 1726776679.20282: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/ /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_file.py && sleep 0' 10792 1726776679.22669: stderr chunk (state=2): >>><<< 10792 1726776679.22679: stdout chunk (state=2): >>><<< 10792 1726776679.22695: _low_level_execute_command() done: rc=0, stdout=, stderr= 10792 1726776679.22700: _low_level_execute_command(): starting 10792 1726776679.22705: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/AnsiballZ_file.py && sleep 0' 10792 1726776679.38690: stdout chunk (state=2): >>> {"path": "/etc/tuned/active_profile", "changed": false, "diff": {"before": {"path": "/etc/tuned/active_profile"}, "after": {"path": "/etc/tuned/active_profile"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/active_profile", "_original_basename": "tmp2swie02d", "recurse": false, "state": "file", "path": "/etc/tuned/active_profile", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10792 1726776679.39827: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10792 1726776679.39877: stderr chunk (state=3): >>><<< 10792 1726776679.39884: stdout chunk (state=3): >>><<< 10792 1726776679.39900: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/active_profile", "changed": false, "diff": {"before": {"path": "/etc/tuned/active_profile"}, "after": {"path": "/etc/tuned/active_profile"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/active_profile", "_original_basename": "tmp2swie02d", "recurse": false, "state": "file", "path": "/etc/tuned/active_profile", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10792 1726776679.39930: done with _execute_module (ansible.legacy.file, {'mode': '0600', 'dest': '/etc/tuned/active_profile', '_original_basename': 'tmp2swie02d', 'recurse': False, 'state': 'file', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10792 1726776679.39943: _low_level_execute_command(): starting 10792 1726776679.39949: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776678.9035478-10792-266422440600387/ > /dev/null 2>&1 && sleep 0' 10792 1726776679.42369: stderr chunk (state=2): >>><<< 10792 1726776679.42377: stdout chunk (state=2): >>><<< 10792 1726776679.42395: _low_level_execute_command() done: rc=0, stdout=, stderr= 10792 1726776679.42403: handler run complete 10792 1726776679.42423: attempt loop complete, returning result 10792 1726776679.42427: _execute() done 10792 1726776679.42432: dumping result to json 10792 1726776679.42438: done dumping result, returning 10792 1726776679.42445: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile [120fa90a-8a95-fee7-0cea-0000000006e7] 10792 1726776679.42451: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e7 10792 1726776679.42485: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e7 10792 1726776679.42489: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "dest": "/etc/tuned/active_profile", "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/etc/tuned/active_profile", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 30, "state": "file", "uid": 0 } 9704 1726776679.42645: no more pending results, returning what we have 9704 1726776679.42648: results queue empty 9704 1726776679.42648: checking for any_errors_fatal 9704 1726776679.42652: done checking for any_errors_fatal 9704 1726776679.42653: checking for max_fail_percentage 9704 1726776679.42654: done checking for max_fail_percentage 9704 1726776679.42655: checking to see if all hosts have failed and the running result is not ok 9704 1726776679.42656: done checking to see if all hosts have failed 9704 1726776679.42656: getting the remaining hosts for this loop 9704 1726776679.42657: done getting the remaining hosts for this loop 9704 1726776679.42660: getting the next task for host managed_node3 9704 1726776679.42666: done getting next task for host managed_node3 9704 1726776679.42671: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual 9704 1726776679.42674: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=9, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776679.42684: getting variables 9704 1726776679.42685: in VariableManager get_vars() 9704 1726776679.42716: Calling all_inventory to load vars for managed_node3 9704 1726776679.42718: Calling groups_inventory to load vars for managed_node3 9704 1726776679.42720: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776679.42731: Calling all_plugins_play to load vars for managed_node3 9704 1726776679.42733: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776679.42736: Calling groups_plugins_play to load vars for managed_node3 9704 1726776679.42890: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776679.43007: done with get_vars() 9704 1726776679.43016: done getting variables 9704 1726776679.43062: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set profile_mode to manual] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:99 Thursday 19 September 2024 16:11:19 -0400 (0:00:00.567) 0:00:32.359 **** 9704 1726776679.43083: entering _queue_task() for managed_node3/copy 9704 1726776679.43243: worker is 1 (out of 1 available) 9704 1726776679.43257: exiting _queue_task() for managed_node3/copy 9704 1726776679.43269: done queuing things up, now waiting for results queue to drain 9704 1726776679.43271: waiting for pending results... 10804 1726776679.43404: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual 10804 1726776679.43522: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e8 10804 1726776679.43540: variable 'ansible_search_path' from source: unknown 10804 1726776679.43545: variable 'ansible_search_path' from source: unknown 10804 1726776679.43576: calling self._execute() 10804 1726776679.43647: variable 'ansible_host' from source: host vars for 'managed_node3' 10804 1726776679.43657: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10804 1726776679.43666: variable 'omit' from source: magic vars 10804 1726776679.43740: variable 'omit' from source: magic vars 10804 1726776679.43780: variable 'omit' from source: magic vars 10804 1726776679.43803: variable '__kernel_settings_tuned_profile_mode' from source: role '' all vars 10804 1726776679.44025: variable '__kernel_settings_tuned_profile_mode' from source: role '' all vars 10804 1726776679.44085: variable '__kernel_settings_tuned_dir' from source: role '' all vars 10804 1726776679.44109: variable 'omit' from source: magic vars 10804 1726776679.44140: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10804 1726776679.44167: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10804 1726776679.44183: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10804 1726776679.44194: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10804 1726776679.44202: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10804 1726776679.44223: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10804 1726776679.44227: variable 'ansible_host' from source: host vars for 'managed_node3' 10804 1726776679.44239: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10804 1726776679.44312: Set connection var ansible_module_compression to ZIP_DEFLATED 10804 1726776679.44320: Set connection var ansible_shell_type to sh 10804 1726776679.44328: Set connection var ansible_timeout to 10 10804 1726776679.44335: Set connection var ansible_pipelining to False 10804 1726776679.44342: Set connection var ansible_shell_executable to /bin/sh 10804 1726776679.44346: Set connection var ansible_connection to ssh 10804 1726776679.44362: variable 'ansible_shell_executable' from source: unknown 10804 1726776679.44366: variable 'ansible_connection' from source: unknown 10804 1726776679.44369: variable 'ansible_module_compression' from source: unknown 10804 1726776679.44372: variable 'ansible_shell_type' from source: unknown 10804 1726776679.44376: variable 'ansible_shell_executable' from source: unknown 10804 1726776679.44379: variable 'ansible_host' from source: host vars for 'managed_node3' 10804 1726776679.44383: variable 'ansible_pipelining' from source: unknown 10804 1726776679.44387: variable 'ansible_timeout' from source: unknown 10804 1726776679.44391: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10804 1726776679.44476: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10804 1726776679.44488: variable 'omit' from source: magic vars 10804 1726776679.44494: starting attempt loop 10804 1726776679.44497: running the handler 10804 1726776679.44508: _low_level_execute_command(): starting 10804 1726776679.44515: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10804 1726776679.46797: stdout chunk (state=2): >>>/root <<< 10804 1726776679.46916: stderr chunk (state=3): >>><<< 10804 1726776679.46923: stdout chunk (state=3): >>><<< 10804 1726776679.46944: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10804 1726776679.46959: _low_level_execute_command(): starting 10804 1726776679.46966: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926 `" && echo ansible-tmp-1726776679.4695206-10804-247835519182926="` echo /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926 `" ) && sleep 0' 10804 1726776679.49441: stdout chunk (state=2): >>>ansible-tmp-1726776679.4695206-10804-247835519182926=/root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926 <<< 10804 1726776679.49565: stderr chunk (state=3): >>><<< 10804 1726776679.49573: stdout chunk (state=3): >>><<< 10804 1726776679.49593: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776679.4695206-10804-247835519182926=/root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926 , stderr= 10804 1726776679.49669: variable 'ansible_module_compression' from source: unknown 10804 1726776679.49718: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10804 1726776679.49750: variable 'ansible_facts' from source: unknown 10804 1726776679.49824: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_stat.py 10804 1726776679.49913: Sending initial data 10804 1726776679.49922: Sent initial data (152 bytes) 10804 1726776679.52382: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpld0u4isu /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_stat.py <<< 10804 1726776679.53325: stderr chunk (state=3): >>><<< 10804 1726776679.53332: stdout chunk (state=3): >>><<< 10804 1726776679.53349: done transferring module to remote 10804 1726776679.53361: _low_level_execute_command(): starting 10804 1726776679.53366: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/ /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_stat.py && sleep 0' 10804 1726776679.55648: stderr chunk (state=2): >>><<< 10804 1726776679.55655: stdout chunk (state=2): >>><<< 10804 1726776679.55668: _low_level_execute_command() done: rc=0, stdout=, stderr= 10804 1726776679.55672: _low_level_execute_command(): starting 10804 1726776679.55677: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_stat.py && sleep 0' 10804 1726776679.71678: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 7, "inode": 148897989, "dev": 51713, "nlink": 1, "atime": 1726776664.3133166, "mtime": 1726776658.1142364, "ctime": 1726776658.1142364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "mimetype": "text/plain", "charset": "us-ascii", "version": "1846195820", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10804 1726776679.72838: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10804 1726776679.72888: stderr chunk (state=3): >>><<< 10804 1726776679.72895: stdout chunk (state=3): >>><<< 10804 1726776679.72915: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 7, "inode": 148897989, "dev": 51713, "nlink": 1, "atime": 1726776664.3133166, "mtime": 1726776658.1142364, "ctime": 1726776658.1142364, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "mimetype": "text/plain", "charset": "us-ascii", "version": "1846195820", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10804 1726776679.72963: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/profile_mode', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10804 1726776679.72998: variable 'ansible_module_compression' from source: unknown 10804 1726776679.73035: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10804 1726776679.73056: variable 'ansible_facts' from source: unknown 10804 1726776679.73113: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_file.py 10804 1726776679.73203: Sending initial data 10804 1726776679.73210: Sent initial data (152 bytes) 10804 1726776679.75737: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp5ryeh71j /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_file.py <<< 10804 1726776679.76767: stderr chunk (state=3): >>><<< 10804 1726776679.76775: stdout chunk (state=3): >>><<< 10804 1726776679.76794: done transferring module to remote 10804 1726776679.76803: _low_level_execute_command(): starting 10804 1726776679.76809: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/ /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_file.py && sleep 0' 10804 1726776679.79118: stderr chunk (state=2): >>><<< 10804 1726776679.79125: stdout chunk (state=2): >>><<< 10804 1726776679.79139: _low_level_execute_command() done: rc=0, stdout=, stderr= 10804 1726776679.79143: _low_level_execute_command(): starting 10804 1726776679.79148: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/AnsiballZ_file.py && sleep 0' 10804 1726776679.95097: stdout chunk (state=2): >>> {"path": "/etc/tuned/profile_mode", "changed": false, "diff": {"before": {"path": "/etc/tuned/profile_mode"}, "after": {"path": "/etc/tuned/profile_mode"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/profile_mode", "_original_basename": "tmpoz7tfk9r", "recurse": false, "state": "file", "path": "/etc/tuned/profile_mode", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10804 1726776679.96218: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10804 1726776679.96269: stderr chunk (state=3): >>><<< 10804 1726776679.96276: stdout chunk (state=3): >>><<< 10804 1726776679.96293: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/profile_mode", "changed": false, "diff": {"before": {"path": "/etc/tuned/profile_mode"}, "after": {"path": "/etc/tuned/profile_mode"}}, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "invocation": {"module_args": {"mode": "0600", "dest": "/etc/tuned/profile_mode", "_original_basename": "tmpoz7tfk9r", "recurse": false, "state": "file", "path": "/etc/tuned/profile_mode", "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10804 1726776679.96322: done with _execute_module (ansible.legacy.file, {'mode': '0600', 'dest': '/etc/tuned/profile_mode', '_original_basename': 'tmpoz7tfk9r', 'recurse': False, 'state': 'file', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10804 1726776679.96335: _low_level_execute_command(): starting 10804 1726776679.96341: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776679.4695206-10804-247835519182926/ > /dev/null 2>&1 && sleep 0' 10804 1726776679.98735: stderr chunk (state=2): >>><<< 10804 1726776679.98742: stdout chunk (state=2): >>><<< 10804 1726776679.98760: _low_level_execute_command() done: rc=0, stdout=, stderr= 10804 1726776679.98768: handler run complete 10804 1726776679.98787: attempt loop complete, returning result 10804 1726776679.98791: _execute() done 10804 1726776679.98794: dumping result to json 10804 1726776679.98800: done dumping result, returning 10804 1726776679.98809: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set profile_mode to manual [120fa90a-8a95-fee7-0cea-0000000006e8] 10804 1726776679.98816: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e8 10804 1726776679.98851: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e8 10804 1726776679.98858: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "dest": "/etc/tuned/profile_mode", "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/etc/tuned/profile_mode", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 7, "state": "file", "uid": 0 } 9704 1726776679.99025: no more pending results, returning what we have 9704 1726776679.99031: results queue empty 9704 1726776679.99032: checking for any_errors_fatal 9704 1726776679.99039: done checking for any_errors_fatal 9704 1726776679.99041: checking for max_fail_percentage 9704 1726776679.99042: done checking for max_fail_percentage 9704 1726776679.99043: checking to see if all hosts have failed and the running result is not ok 9704 1726776679.99043: done checking to see if all hosts have failed 9704 1726776679.99044: getting the remaining hosts for this loop 9704 1726776679.99045: done getting the remaining hosts for this loop 9704 1726776679.99048: getting the next task for host managed_node3 9704 1726776679.99054: done getting next task for host managed_node3 9704 1726776679.99057: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get current config 9704 1726776679.99060: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=10, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776679.99070: getting variables 9704 1726776679.99071: in VariableManager get_vars() 9704 1726776679.99105: Calling all_inventory to load vars for managed_node3 9704 1726776679.99107: Calling groups_inventory to load vars for managed_node3 9704 1726776679.99109: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776679.99119: Calling all_plugins_play to load vars for managed_node3 9704 1726776679.99121: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776679.99124: Calling groups_plugins_play to load vars for managed_node3 9704 1726776679.99242: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776679.99373: done with get_vars() 9704 1726776679.99381: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Get current config] ********** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:107 Thursday 19 September 2024 16:11:19 -0400 (0:00:00.563) 0:00:32.923 **** 9704 1726776679.99443: entering _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776679.99609: worker is 1 (out of 1 available) 9704 1726776679.99624: exiting _queue_task() for managed_node3/fedora.linux_system_roles.kernel_settings_get_config 9704 1726776679.99638: done queuing things up, now waiting for results queue to drain 9704 1726776679.99641: waiting for pending results... 10819 1726776679.99769: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get current config 10819 1726776679.99891: in run() - task 120fa90a-8a95-fee7-0cea-0000000006e9 10819 1726776679.99907: variable 'ansible_search_path' from source: unknown 10819 1726776679.99912: variable 'ansible_search_path' from source: unknown 10819 1726776679.99941: calling self._execute() 10819 1726776680.00014: variable 'ansible_host' from source: host vars for 'managed_node3' 10819 1726776680.00022: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10819 1726776680.00033: variable 'omit' from source: magic vars 10819 1726776680.00110: variable 'omit' from source: magic vars 10819 1726776680.00151: variable 'omit' from source: magic vars 10819 1726776680.00174: variable '__kernel_settings_profile_filename' from source: role '' all vars 10819 1726776680.00392: variable '__kernel_settings_profile_filename' from source: role '' all vars 10819 1726776680.00458: variable '__kernel_settings_profile_dir' from source: role '' all vars 10819 1726776680.00583: variable '__kernel_settings_profile_parent' from source: set_fact 10819 1726776680.00591: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10819 1726776680.00625: variable 'omit' from source: magic vars 10819 1726776680.00661: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10819 1726776680.00687: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10819 1726776680.00706: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10819 1726776680.00719: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10819 1726776680.00731: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10819 1726776680.00754: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10819 1726776680.00762: variable 'ansible_host' from source: host vars for 'managed_node3' 10819 1726776680.00767: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10819 1726776680.00833: Set connection var ansible_module_compression to ZIP_DEFLATED 10819 1726776680.00841: Set connection var ansible_shell_type to sh 10819 1726776680.00848: Set connection var ansible_timeout to 10 10819 1726776680.00851: Set connection var ansible_pipelining to False 10819 1726776680.00858: Set connection var ansible_shell_executable to /bin/sh 10819 1726776680.00860: Set connection var ansible_connection to ssh 10819 1726776680.00874: variable 'ansible_shell_executable' from source: unknown 10819 1726776680.00877: variable 'ansible_connection' from source: unknown 10819 1726776680.00879: variable 'ansible_module_compression' from source: unknown 10819 1726776680.00880: variable 'ansible_shell_type' from source: unknown 10819 1726776680.00882: variable 'ansible_shell_executable' from source: unknown 10819 1726776680.00883: variable 'ansible_host' from source: host vars for 'managed_node3' 10819 1726776680.00885: variable 'ansible_pipelining' from source: unknown 10819 1726776680.00887: variable 'ansible_timeout' from source: unknown 10819 1726776680.00889: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10819 1726776680.01011: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10819 1726776680.01020: variable 'omit' from source: magic vars 10819 1726776680.01024: starting attempt loop 10819 1726776680.01026: running the handler 10819 1726776680.01040: _low_level_execute_command(): starting 10819 1726776680.01049: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10819 1726776680.03366: stdout chunk (state=2): >>>/root <<< 10819 1726776680.03487: stderr chunk (state=3): >>><<< 10819 1726776680.03493: stdout chunk (state=3): >>><<< 10819 1726776680.03509: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10819 1726776680.03522: _low_level_execute_command(): starting 10819 1726776680.03528: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556 `" && echo ansible-tmp-1726776680.0351648-10819-165103890888556="` echo /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556 `" ) && sleep 0' 10819 1726776680.05938: stdout chunk (state=2): >>>ansible-tmp-1726776680.0351648-10819-165103890888556=/root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556 <<< 10819 1726776680.06062: stderr chunk (state=3): >>><<< 10819 1726776680.06068: stdout chunk (state=3): >>><<< 10819 1726776680.06081: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776680.0351648-10819-165103890888556=/root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556 , stderr= 10819 1726776680.06115: variable 'ansible_module_compression' from source: unknown 10819 1726776680.06145: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible_collections.fedora.linux_system_roles.plugins.modules.kernel_settings_get_config-ZIP_DEFLATED 10819 1726776680.06174: variable 'ansible_facts' from source: unknown 10819 1726776680.06243: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/AnsiballZ_kernel_settings_get_config.py 10819 1726776680.06335: Sending initial data 10819 1726776680.06342: Sent initial data (174 bytes) 10819 1726776680.08785: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp6x937pyh /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/AnsiballZ_kernel_settings_get_config.py <<< 10819 1726776680.09774: stderr chunk (state=3): >>><<< 10819 1726776680.09785: stdout chunk (state=3): >>><<< 10819 1726776680.09806: done transferring module to remote 10819 1726776680.09819: _low_level_execute_command(): starting 10819 1726776680.09825: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/ /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10819 1726776680.12175: stderr chunk (state=2): >>><<< 10819 1726776680.12184: stdout chunk (state=2): >>><<< 10819 1726776680.12198: _low_level_execute_command() done: rc=0, stdout=, stderr= 10819 1726776680.12203: _low_level_execute_command(): starting 10819 1726776680.12208: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/AnsiballZ_kernel_settings_get_config.py && sleep 0' 10819 1726776680.27804: stdout chunk (state=2): >>> {"changed": false, "data": {"main": {"summary": "kernel settings"}, "sysctl": {"fs.epoll.max_user_watches": "785592", "fs.file-max": "379724"}, "sysfs": {"/sys/kernel/debug/x86/ibrs_enabled": "0", "/sys/kernel/debug/x86/pti_enabled": "0", "/sys/kernel/debug/x86/retp_enabled": "0"}, "vm": {"transparent_hugepages": "madvise"}}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}} <<< 10819 1726776680.28902: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10819 1726776680.28951: stderr chunk (state=3): >>><<< 10819 1726776680.28959: stdout chunk (state=3): >>><<< 10819 1726776680.28977: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "data": {"main": {"summary": "kernel settings"}, "sysctl": {"fs.epoll.max_user_watches": "785592", "fs.file-max": "379724"}, "sysfs": {"/sys/kernel/debug/x86/ibrs_enabled": "0", "/sys/kernel/debug/x86/pti_enabled": "0", "/sys/kernel/debug/x86/retp_enabled": "0"}, "vm": {"transparent_hugepages": "madvise"}}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf"}}} , stderr=Shared connection to 10.31.8.186 closed. 10819 1726776680.29004: done with _execute_module (fedora.linux_system_roles.kernel_settings_get_config, {'path': '/etc/tuned/kernel_settings/tuned.conf', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'fedora.linux_system_roles.kernel_settings_get_config', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10819 1726776680.29017: _low_level_execute_command(): starting 10819 1726776680.29023: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776680.0351648-10819-165103890888556/ > /dev/null 2>&1 && sleep 0' 10819 1726776680.31445: stderr chunk (state=2): >>><<< 10819 1726776680.31454: stdout chunk (state=2): >>><<< 10819 1726776680.31470: _low_level_execute_command() done: rc=0, stdout=, stderr= 10819 1726776680.31477: handler run complete 10819 1726776680.31491: attempt loop complete, returning result 10819 1726776680.31495: _execute() done 10819 1726776680.31499: dumping result to json 10819 1726776680.31503: done dumping result, returning 10819 1726776680.31510: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get current config [120fa90a-8a95-fee7-0cea-0000000006e9] 10819 1726776680.31516: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e9 10819 1726776680.31547: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006e9 10819 1726776680.31551: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "data": { "main": { "summary": "kernel settings" }, "sysctl": { "fs.epoll.max_user_watches": "785592", "fs.file-max": "379724" }, "sysfs": { "/sys/kernel/debug/x86/ibrs_enabled": "0", "/sys/kernel/debug/x86/pti_enabled": "0", "/sys/kernel/debug/x86/retp_enabled": "0" }, "vm": { "transparent_hugepages": "madvise" } } } 9704 1726776680.31712: no more pending results, returning what we have 9704 1726776680.31716: results queue empty 9704 1726776680.31717: checking for any_errors_fatal 9704 1726776680.31723: done checking for any_errors_fatal 9704 1726776680.31724: checking for max_fail_percentage 9704 1726776680.31725: done checking for max_fail_percentage 9704 1726776680.31726: checking to see if all hosts have failed and the running result is not ok 9704 1726776680.31726: done checking to see if all hosts have failed 9704 1726776680.31727: getting the remaining hosts for this loop 9704 1726776680.31730: done getting the remaining hosts for this loop 9704 1726776680.31733: getting the next task for host managed_node3 9704 1726776680.31741: done getting next task for host managed_node3 9704 1726776680.31744: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings 9704 1726776680.31748: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=11, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776680.31759: getting variables 9704 1726776680.31761: in VariableManager get_vars() 9704 1726776680.31796: Calling all_inventory to load vars for managed_node3 9704 1726776680.31798: Calling groups_inventory to load vars for managed_node3 9704 1726776680.31800: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776680.31809: Calling all_plugins_play to load vars for managed_node3 9704 1726776680.31811: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776680.31814: Calling groups_plugins_play to load vars for managed_node3 9704 1726776680.31973: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776680.32100: done with get_vars() 9704 1726776680.32108: done getting variables 9704 1726776680.32154: Loading ActionModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/template.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Apply kernel settings] ******* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:112 Thursday 19 September 2024 16:11:20 -0400 (0:00:00.327) 0:00:33.250 **** 9704 1726776680.32179: entering _queue_task() for managed_node3/template 9704 1726776680.32342: worker is 1 (out of 1 available) 9704 1726776680.32360: exiting _queue_task() for managed_node3/template 9704 1726776680.32372: done queuing things up, now waiting for results queue to drain 9704 1726776680.32374: waiting for pending results... 10827 1726776680.32500: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings 10827 1726776680.32618: in run() - task 120fa90a-8a95-fee7-0cea-0000000006ea 10827 1726776680.32636: variable 'ansible_search_path' from source: unknown 10827 1726776680.32641: variable 'ansible_search_path' from source: unknown 10827 1726776680.32669: calling self._execute() 10827 1726776680.32739: variable 'ansible_host' from source: host vars for 'managed_node3' 10827 1726776680.32748: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10827 1726776680.32757: variable 'omit' from source: magic vars 10827 1726776680.32832: variable 'omit' from source: magic vars 10827 1726776680.32871: variable 'omit' from source: magic vars 10827 1726776680.33104: variable '__kernel_settings_profile_src' from source: role '' all vars 10827 1726776680.33113: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10827 1726776680.33170: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10827 1726776680.33190: variable '__kernel_settings_profile_filename' from source: role '' all vars 10827 1726776680.33238: variable '__kernel_settings_profile_filename' from source: role '' all vars 10827 1726776680.33288: variable '__kernel_settings_profile_dir' from source: role '' all vars 10827 1726776680.33350: variable '__kernel_settings_profile_parent' from source: set_fact 10827 1726776680.33358: variable '__kernel_settings_tuned_profile' from source: role '' all vars 10827 1726776680.33381: variable 'omit' from source: magic vars 10827 1726776680.33410: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10827 1726776680.33440: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10827 1726776680.33459: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10827 1726776680.33473: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10827 1726776680.33484: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10827 1726776680.33507: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10827 1726776680.33511: variable 'ansible_host' from source: host vars for 'managed_node3' 10827 1726776680.33515: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10827 1726776680.33583: Set connection var ansible_module_compression to ZIP_DEFLATED 10827 1726776680.33591: Set connection var ansible_shell_type to sh 10827 1726776680.33598: Set connection var ansible_timeout to 10 10827 1726776680.33604: Set connection var ansible_pipelining to False 10827 1726776680.33611: Set connection var ansible_shell_executable to /bin/sh 10827 1726776680.33614: Set connection var ansible_connection to ssh 10827 1726776680.33632: variable 'ansible_shell_executable' from source: unknown 10827 1726776680.33636: variable 'ansible_connection' from source: unknown 10827 1726776680.33640: variable 'ansible_module_compression' from source: unknown 10827 1726776680.33643: variable 'ansible_shell_type' from source: unknown 10827 1726776680.33646: variable 'ansible_shell_executable' from source: unknown 10827 1726776680.33649: variable 'ansible_host' from source: host vars for 'managed_node3' 10827 1726776680.33653: variable 'ansible_pipelining' from source: unknown 10827 1726776680.33656: variable 'ansible_timeout' from source: unknown 10827 1726776680.33660: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10827 1726776680.33748: Loading ActionModule 'template' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/template.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10827 1726776680.33761: variable 'omit' from source: magic vars 10827 1726776680.33767: starting attempt loop 10827 1726776680.33771: running the handler 10827 1726776680.33782: _low_level_execute_command(): starting 10827 1726776680.33789: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10827 1726776680.36086: stdout chunk (state=2): >>>/root <<< 10827 1726776680.36205: stderr chunk (state=3): >>><<< 10827 1726776680.36211: stdout chunk (state=3): >>><<< 10827 1726776680.36230: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10827 1726776680.36243: _low_level_execute_command(): starting 10827 1726776680.36250: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743 `" && echo ansible-tmp-1726776680.3623786-10827-258690884258743="` echo /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743 `" ) && sleep 0' 10827 1726776680.38681: stdout chunk (state=2): >>>ansible-tmp-1726776680.3623786-10827-258690884258743=/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743 <<< 10827 1726776680.38808: stderr chunk (state=3): >>><<< 10827 1726776680.38815: stdout chunk (state=3): >>><<< 10827 1726776680.38828: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776680.3623786-10827-258690884258743=/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743 , stderr= 10827 1726776680.38845: evaluation_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks 10827 1726776680.38863: search_path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/templates/kernel_settings.j2 /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/kernel_settings.j2 10827 1726776680.38883: variable 'ansible_search_path' from source: unknown 10827 1726776680.39458: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10827 1726776680.40892: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10827 1726776680.40947: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10827 1726776680.40978: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10827 1726776680.41006: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10827 1726776680.41027: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10827 1726776680.41209: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10827 1726776680.41234: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10827 1726776680.41259: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10827 1726776680.41286: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10827 1726776680.41298: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10827 1726776680.41520: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10827 1726776680.41541: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10827 1726776680.41563: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10827 1726776680.41590: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10827 1726776680.41601: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10827 1726776680.41846: variable 'ansible_managed' from source: unknown 10827 1726776680.41853: variable '__sections' from source: task vars 10827 1726776680.41941: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10827 1726776680.41959: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10827 1726776680.41977: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10827 1726776680.42005: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10827 1726776680.42016: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10827 1726776680.42085: variable 'kernel_settings_sysctl' from source: include params 10827 1726776680.42092: variable '__kernel_settings_state_empty' from source: role '' all vars 10827 1726776680.42099: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10827 1726776680.42130: variable '__sysctl_old' from source: task vars 10827 1726776680.42174: variable '__sysctl_old' from source: task vars 10827 1726776680.42310: variable 'kernel_settings_purge' from source: include params 10827 1726776680.42317: variable 'kernel_settings_sysctl' from source: include params 10827 1726776680.42321: variable '__kernel_settings_state_empty' from source: role '' all vars 10827 1726776680.42327: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10827 1726776680.42334: variable '__kernel_settings_profile_contents' from source: set_fact 10827 1726776680.42459: variable 'kernel_settings_sysfs' from source: include params 10827 1726776680.42466: variable '__kernel_settings_state_empty' from source: role '' all vars 10827 1726776680.42472: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10827 1726776680.42485: variable '__sysfs_old' from source: task vars 10827 1726776680.42524: variable '__sysfs_old' from source: task vars 10827 1726776680.42680: variable 'kernel_settings_purge' from source: include params 10827 1726776680.42687: variable 'kernel_settings_sysfs' from source: include params 10827 1726776680.42692: variable '__kernel_settings_state_empty' from source: role '' all vars 10827 1726776680.42698: variable '__kernel_settings_previous_replaced' from source: role '' all vars 10827 1726776680.42702: variable '__kernel_settings_profile_contents' from source: set_fact 10827 1726776680.42732: variable 'kernel_settings_systemd_cpu_affinity' from source: include params 10827 1726776680.42741: variable '__systemd_old' from source: task vars 10827 1726776680.42783: variable '__systemd_old' from source: task vars 10827 1726776680.42936: variable 'kernel_settings_purge' from source: include params 10827 1726776680.42943: variable 'kernel_settings_systemd_cpu_affinity' from source: include params 10827 1726776680.42948: variable '__kernel_settings_state_absent' from source: role '' all vars 10827 1726776680.42953: variable '__kernel_settings_profile_contents' from source: set_fact 10827 1726776680.42969: variable 'kernel_settings_transparent_hugepages' from source: include params 10827 1726776680.42975: variable 'kernel_settings_transparent_hugepages_defrag' from source: include params 10827 1726776680.42980: variable '__trans_huge_old' from source: task vars 10827 1726776680.43026: variable '__trans_huge_old' from source: task vars 10827 1726776680.43176: variable 'kernel_settings_purge' from source: include params 10827 1726776680.43183: variable 'kernel_settings_transparent_hugepages' from source: include params 10827 1726776680.43188: variable '__kernel_settings_state_absent' from source: role '' all vars 10827 1726776680.43193: variable '__kernel_settings_profile_contents' from source: set_fact 10827 1726776680.43210: variable '__trans_defrag_old' from source: task vars 10827 1726776680.43259: variable '__trans_defrag_old' from source: task vars 10827 1726776680.43402: variable 'kernel_settings_purge' from source: include params 10827 1726776680.43409: variable 'kernel_settings_transparent_hugepages_defrag' from source: include params 10827 1726776680.43414: variable '__kernel_settings_state_absent' from source: role '' all vars 10827 1726776680.43420: variable '__kernel_settings_profile_contents' from source: set_fact 10827 1726776680.43443: variable '__kernel_settings_state_absent' from source: role '' all vars 10827 1726776680.43458: variable '__kernel_settings_state_absent' from source: role '' all vars 10827 1726776680.43469: variable '__kernel_settings_state_absent' from source: role '' all vars 10827 1726776680.44133: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10827 1726776680.44179: variable 'ansible_module_compression' from source: unknown 10827 1726776680.44220: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10827 1726776680.44241: variable 'ansible_facts' from source: unknown 10827 1726776680.44302: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_stat.py 10827 1726776680.44392: Sending initial data 10827 1726776680.44399: Sent initial data (152 bytes) 10827 1726776680.46935: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpnp6jdalr /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_stat.py <<< 10827 1726776680.47907: stderr chunk (state=3): >>><<< 10827 1726776680.47914: stdout chunk (state=3): >>><<< 10827 1726776680.47933: done transferring module to remote 10827 1726776680.47944: _low_level_execute_command(): starting 10827 1726776680.47949: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/ /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_stat.py && sleep 0' 10827 1726776680.50297: stderr chunk (state=2): >>><<< 10827 1726776680.50304: stdout chunk (state=2): >>><<< 10827 1726776680.50318: _low_level_execute_command() done: rc=0, stdout=, stderr= 10827 1726776680.50322: _low_level_execute_command(): starting 10827 1726776680.50327: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_stat.py && sleep 0' 10827 1726776680.66501: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 312, "inode": 159383749, "dev": 51713, "nlink": 1, "atime": 1726776658.092236, "mtime": 1726776657.1702242, "ctime": 1726776657.4132273, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": true, "xgrp": false, "woth": false, "roth": true, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "mimetype": "text/plain", "charset": "us-ascii", "version": "646792945", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10827 1726776680.67630: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10827 1726776680.67678: stderr chunk (state=3): >>><<< 10827 1726776680.67684: stdout chunk (state=3): >>><<< 10827 1726776680.67700: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 312, "inode": 159383749, "dev": 51713, "nlink": 1, "atime": 1726776658.092236, "mtime": 1726776657.1702242, "ctime": 1726776657.4132273, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": true, "xgrp": false, "woth": false, "roth": true, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "ba15904bb90578344fad097ce2f46f9231275eae", "mimetype": "text/plain", "charset": "us-ascii", "version": "646792945", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings/tuned.conf", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10827 1726776680.67740: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/kernel_settings/tuned.conf', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10827 1726776680.67824: Sending initial data 10827 1726776680.67833: Sent initial data (160 bytes) 10827 1726776680.70317: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp2eafv1ft/kernel_settings.j2 /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source <<< 10827 1726776680.70569: stderr chunk (state=3): >>><<< 10827 1726776680.70575: stdout chunk (state=3): >>><<< 10827 1726776680.70588: _low_level_execute_command(): starting 10827 1726776680.70594: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/ /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source && sleep 0' 10827 1726776680.72887: stderr chunk (state=2): >>><<< 10827 1726776680.72893: stdout chunk (state=2): >>><<< 10827 1726776680.72905: _low_level_execute_command() done: rc=0, stdout=, stderr= 10827 1726776680.72923: variable 'ansible_module_compression' from source: unknown 10827 1726776680.72956: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.copy-ZIP_DEFLATED 10827 1726776680.72976: variable 'ansible_facts' from source: unknown 10827 1726776680.73032: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_copy.py 10827 1726776680.73109: Sending initial data 10827 1726776680.73116: Sent initial data (152 bytes) 10827 1726776680.75568: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpptm1hcqx /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_copy.py <<< 10827 1726776680.76565: stderr chunk (state=3): >>><<< 10827 1726776680.76574: stdout chunk (state=3): >>><<< 10827 1726776680.76594: done transferring module to remote 10827 1726776680.76604: _low_level_execute_command(): starting 10827 1726776680.76609: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/ /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_copy.py && sleep 0' 10827 1726776680.78983: stderr chunk (state=2): >>><<< 10827 1726776680.78993: stdout chunk (state=2): >>><<< 10827 1726776680.79007: _low_level_execute_command() done: rc=0, stdout=, stderr= 10827 1726776680.79012: _low_level_execute_command(): starting 10827 1726776680.79018: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/AnsiballZ_copy.py && sleep 0' 10827 1726776680.95289: stdout chunk (state=2): >>> {"dest": "/etc/tuned/kernel_settings/tuned.conf", "src": "/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source", "md5sum": "7d83891795eeb6debeff7e2812501630", "checksum": "e44ba7fc7046252a1b6772f7347d0e7b9b48a069", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0644", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 86, "invocation": {"module_args": {"src": "/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source", "dest": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "follow": false, "_original_basename": "kernel_settings.j2", "checksum": "e44ba7fc7046252a1b6772f7347d0e7b9b48a069", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10827 1726776680.96453: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10827 1726776680.96505: stderr chunk (state=3): >>><<< 10827 1726776680.96511: stdout chunk (state=3): >>><<< 10827 1726776680.96527: _low_level_execute_command() done: rc=0, stdout= {"dest": "/etc/tuned/kernel_settings/tuned.conf", "src": "/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source", "md5sum": "7d83891795eeb6debeff7e2812501630", "checksum": "e44ba7fc7046252a1b6772f7347d0e7b9b48a069", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0644", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 86, "invocation": {"module_args": {"src": "/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source", "dest": "/etc/tuned/kernel_settings/tuned.conf", "mode": "0644", "follow": false, "_original_basename": "kernel_settings.j2", "checksum": "e44ba7fc7046252a1b6772f7347d0e7b9b48a069", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10827 1726776680.96555: done with _execute_module (ansible.legacy.copy, {'src': '/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source', 'dest': '/etc/tuned/kernel_settings/tuned.conf', 'mode': '0644', 'follow': False, '_original_basename': 'kernel_settings.j2', 'checksum': 'e44ba7fc7046252a1b6772f7347d0e7b9b48a069', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10827 1726776680.96586: _low_level_execute_command(): starting 10827 1726776680.96594: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/ > /dev/null 2>&1 && sleep 0' 10827 1726776680.98983: stderr chunk (state=2): >>><<< 10827 1726776680.98993: stdout chunk (state=2): >>><<< 10827 1726776680.99008: _low_level_execute_command() done: rc=0, stdout=, stderr= 10827 1726776680.99021: handler run complete 10827 1726776680.99043: attempt loop complete, returning result 10827 1726776680.99047: _execute() done 10827 1726776680.99051: dumping result to json 10827 1726776680.99060: done dumping result, returning 10827 1726776680.99068: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Apply kernel settings [120fa90a-8a95-fee7-0cea-0000000006ea] 10827 1726776680.99074: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ea 10827 1726776680.99117: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ea 10827 1726776680.99121: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "e44ba7fc7046252a1b6772f7347d0e7b9b48a069", "dest": "/etc/tuned/kernel_settings/tuned.conf", "gid": 0, "group": "root", "md5sum": "7d83891795eeb6debeff7e2812501630", "mode": "0644", "owner": "root", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 86, "src": "/root/.ansible/tmp/ansible-tmp-1726776680.3623786-10827-258690884258743/source", "state": "file", "uid": 0 } 9704 1726776680.99304: no more pending results, returning what we have 9704 1726776680.99308: results queue empty 9704 1726776680.99309: checking for any_errors_fatal 9704 1726776680.99317: done checking for any_errors_fatal 9704 1726776680.99317: checking for max_fail_percentage 9704 1726776680.99318: done checking for max_fail_percentage 9704 1726776680.99319: checking to see if all hosts have failed and the running result is not ok 9704 1726776680.99320: done checking to see if all hosts have failed 9704 1726776680.99320: getting the remaining hosts for this loop 9704 1726776680.99321: done getting the remaining hosts for this loop 9704 1726776680.99324: getting the next task for host managed_node3 9704 1726776680.99332: done getting next task for host managed_node3 9704 1726776680.99336: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes 9704 1726776680.99339: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=12, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776680.99349: getting variables 9704 1726776680.99350: in VariableManager get_vars() 9704 1726776680.99383: Calling all_inventory to load vars for managed_node3 9704 1726776680.99386: Calling groups_inventory to load vars for managed_node3 9704 1726776680.99388: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776680.99397: Calling all_plugins_play to load vars for managed_node3 9704 1726776680.99399: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776680.99401: Calling groups_plugins_play to load vars for managed_node3 9704 1726776680.99516: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776680.99647: done with get_vars() 9704 1726776680.99656: done getting variables 9704 1726776680.99718: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:149 Thursday 19 September 2024 16:11:20 -0400 (0:00:00.675) 0:00:33.926 **** 9704 1726776680.99756: entering _queue_task() for managed_node3/service 9704 1726776680.99956: worker is 1 (out of 1 available) 9704 1726776680.99970: exiting _queue_task() for managed_node3/service 9704 1726776680.99981: done queuing things up, now waiting for results queue to drain 9704 1726776680.99983: waiting for pending results... 10846 1726776681.00305: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes 10846 1726776681.00456: in run() - task 120fa90a-8a95-fee7-0cea-0000000006eb 10846 1726776681.00473: variable 'ansible_search_path' from source: unknown 10846 1726776681.00478: variable 'ansible_search_path' from source: unknown 10846 1726776681.00519: variable '__kernel_settings_services' from source: include_vars 10846 1726776681.00914: variable '__kernel_settings_services' from source: include_vars 10846 1726776681.00987: variable 'omit' from source: magic vars 10846 1726776681.01122: variable 'ansible_host' from source: host vars for 'managed_node3' 10846 1726776681.01136: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10846 1726776681.01149: variable 'omit' from source: magic vars 10846 1726776681.01410: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10846 1726776681.01652: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10846 1726776681.01696: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10846 1726776681.01731: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10846 1726776681.01765: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10846 1726776681.01872: variable '__kernel_settings_register_profile' from source: set_fact 10846 1726776681.01886: variable '__kernel_settings_register_mode' from source: set_fact 10846 1726776681.01906: Evaluated conditional (__kernel_settings_register_profile is changed or __kernel_settings_register_mode is changed): False 10846 1726776681.01911: when evaluation is False, skipping this task 10846 1726776681.01944: variable 'item' from source: unknown 10846 1726776681.02009: variable 'item' from source: unknown skipping: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": false, "false_condition": "__kernel_settings_register_profile is changed or __kernel_settings_register_mode is changed", "item": "tuned", "skip_reason": "Conditional result was False" } 10846 1726776681.02045: dumping result to json 10846 1726776681.02051: done dumping result, returning 10846 1726776681.02057: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes [120fa90a-8a95-fee7-0cea-0000000006eb] 10846 1726776681.02063: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006eb 10846 1726776681.02086: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006eb 10846 1726776681.02090: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false } MSG: All items skipped 9704 1726776681.02398: no more pending results, returning what we have 9704 1726776681.02402: results queue empty 9704 1726776681.02403: checking for any_errors_fatal 9704 1726776681.02418: done checking for any_errors_fatal 9704 1726776681.02419: checking for max_fail_percentage 9704 1726776681.02421: done checking for max_fail_percentage 9704 1726776681.02422: checking to see if all hosts have failed and the running result is not ok 9704 1726776681.02422: done checking to see if all hosts have failed 9704 1726776681.02423: getting the remaining hosts for this loop 9704 1726776681.02425: done getting the remaining hosts for this loop 9704 1726776681.02430: getting the next task for host managed_node3 9704 1726776681.02438: done getting next task for host managed_node3 9704 1726776681.02442: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings 9704 1726776681.02446: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=13, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776681.02461: getting variables 9704 1726776681.02463: in VariableManager get_vars() 9704 1726776681.02504: Calling all_inventory to load vars for managed_node3 9704 1726776681.02507: Calling groups_inventory to load vars for managed_node3 9704 1726776681.02509: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776681.02522: Calling all_plugins_play to load vars for managed_node3 9704 1726776681.02525: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776681.02530: Calling groups_plugins_play to load vars for managed_node3 9704 1726776681.02709: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776681.02919: done with get_vars() 9704 1726776681.02932: done getting variables 9704 1726776681.02995: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Tuned apply settings] ******** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:157 Thursday 19 September 2024 16:11:21 -0400 (0:00:00.032) 0:00:33.959 **** 9704 1726776681.03041: entering _queue_task() for managed_node3/command 9704 1726776681.03472: worker is 1 (out of 1 available) 9704 1726776681.03484: exiting _queue_task() for managed_node3/command 9704 1726776681.03495: done queuing things up, now waiting for results queue to drain 9704 1726776681.03497: waiting for pending results... 10848 1726776681.03702: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings 10848 1726776681.03839: in run() - task 120fa90a-8a95-fee7-0cea-0000000006ec 10848 1726776681.03855: variable 'ansible_search_path' from source: unknown 10848 1726776681.03858: variable 'ansible_search_path' from source: unknown 10848 1726776681.03888: calling self._execute() 10848 1726776681.03967: variable 'ansible_host' from source: host vars for 'managed_node3' 10848 1726776681.03975: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10848 1726776681.03983: variable 'omit' from source: magic vars 10848 1726776681.04415: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10848 1726776681.04824: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10848 1726776681.04867: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10848 1726776681.04899: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10848 1726776681.04932: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10848 1726776681.05033: variable '__kernel_settings_register_profile' from source: set_fact 10848 1726776681.05060: Evaluated conditional (not __kernel_settings_register_profile is changed): True 10848 1726776681.05176: variable '__kernel_settings_register_mode' from source: set_fact 10848 1726776681.05188: Evaluated conditional (not __kernel_settings_register_mode is changed): True 10848 1726776681.05295: variable '__kernel_settings_register_apply' from source: set_fact 10848 1726776681.05306: Evaluated conditional (__kernel_settings_register_apply is changed): True 10848 1726776681.05312: variable 'omit' from source: magic vars 10848 1726776681.05363: variable 'omit' from source: magic vars 10848 1726776681.05486: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10848 1726776681.07511: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10848 1726776681.07581: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10848 1726776681.07614: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10848 1726776681.07648: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10848 1726776681.07673: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10848 1726776681.07742: variable '__kernel_settings_active_profile' from source: set_fact 10848 1726776681.07774: variable 'omit' from source: magic vars 10848 1726776681.07801: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10848 1726776681.07828: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10848 1726776681.07873: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10848 1726776681.07891: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10848 1726776681.07902: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10848 1726776681.07931: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10848 1726776681.07937: variable 'ansible_host' from source: host vars for 'managed_node3' 10848 1726776681.07943: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10848 1726776681.08036: Set connection var ansible_module_compression to ZIP_DEFLATED 10848 1726776681.08045: Set connection var ansible_shell_type to sh 10848 1726776681.08054: Set connection var ansible_timeout to 10 10848 1726776681.08060: Set connection var ansible_pipelining to False 10848 1726776681.08070: Set connection var ansible_shell_executable to /bin/sh 10848 1726776681.08074: Set connection var ansible_connection to ssh 10848 1726776681.08094: variable 'ansible_shell_executable' from source: unknown 10848 1726776681.08099: variable 'ansible_connection' from source: unknown 10848 1726776681.08102: variable 'ansible_module_compression' from source: unknown 10848 1726776681.08105: variable 'ansible_shell_type' from source: unknown 10848 1726776681.08108: variable 'ansible_shell_executable' from source: unknown 10848 1726776681.08111: variable 'ansible_host' from source: host vars for 'managed_node3' 10848 1726776681.08114: variable 'ansible_pipelining' from source: unknown 10848 1726776681.08117: variable 'ansible_timeout' from source: unknown 10848 1726776681.08121: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10848 1726776681.08201: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10848 1726776681.08212: variable 'omit' from source: magic vars 10848 1726776681.08217: starting attempt loop 10848 1726776681.08220: running the handler 10848 1726776681.08233: _low_level_execute_command(): starting 10848 1726776681.08238: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10848 1726776681.10959: stdout chunk (state=2): >>>/root <<< 10848 1726776681.11098: stderr chunk (state=3): >>><<< 10848 1726776681.11106: stdout chunk (state=3): >>><<< 10848 1726776681.11124: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10848 1726776681.11139: _low_level_execute_command(): starting 10848 1726776681.11146: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669 `" && echo ansible-tmp-1726776681.1113408-10848-238830726360669="` echo /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669 `" ) && sleep 0' 10848 1726776681.14246: stdout chunk (state=2): >>>ansible-tmp-1726776681.1113408-10848-238830726360669=/root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669 <<< 10848 1726776681.14396: stderr chunk (state=3): >>><<< 10848 1726776681.14404: stdout chunk (state=3): >>><<< 10848 1726776681.14420: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776681.1113408-10848-238830726360669=/root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669 , stderr= 10848 1726776681.14450: variable 'ansible_module_compression' from source: unknown 10848 1726776681.14503: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 10848 1726776681.14538: variable 'ansible_facts' from source: unknown 10848 1726776681.14645: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/AnsiballZ_command.py 10848 1726776681.15134: Sending initial data 10848 1726776681.15141: Sent initial data (155 bytes) 10848 1726776681.17652: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpyypb_d8f /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/AnsiballZ_command.py <<< 10848 1726776681.18818: stderr chunk (state=3): >>><<< 10848 1726776681.18824: stdout chunk (state=3): >>><<< 10848 1726776681.18841: done transferring module to remote 10848 1726776681.18851: _low_level_execute_command(): starting 10848 1726776681.18854: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/ /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/AnsiballZ_command.py && sleep 0' 10848 1726776681.21182: stderr chunk (state=2): >>><<< 10848 1726776681.21189: stdout chunk (state=2): >>><<< 10848 1726776681.21202: _low_level_execute_command() done: rc=0, stdout=, stderr= 10848 1726776681.21207: _low_level_execute_command(): starting 10848 1726776681.21211: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/AnsiballZ_command.py && sleep 0' 10848 1726776682.53855: stdout chunk (state=2): >>> {"changed": true, "stdout": "", "stderr": "", "rc": 0, "cmd": ["tuned-adm", "profile", "virtual-guest kernel_settings"], "start": "2024-09-19 16:11:21.363370", "end": "2024-09-19 16:11:22.533945", "delta": "0:00:01.170575", "msg": "", "invocation": {"module_args": {"_raw_params": "tuned-adm profile 'virtual-guest kernel_settings'", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 10848 1726776682.54786: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10848 1726776682.54839: stderr chunk (state=3): >>><<< 10848 1726776682.54845: stdout chunk (state=3): >>><<< 10848 1726776682.54864: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "", "rc": 0, "cmd": ["tuned-adm", "profile", "virtual-guest kernel_settings"], "start": "2024-09-19 16:11:21.363370", "end": "2024-09-19 16:11:22.533945", "delta": "0:00:01.170575", "msg": "", "invocation": {"module_args": {"_raw_params": "tuned-adm profile 'virtual-guest kernel_settings'", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10848 1726776682.54891: done with _execute_module (ansible.legacy.command, {'_raw_params': "tuned-adm profile 'virtual-guest kernel_settings'", '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10848 1726776682.54900: _low_level_execute_command(): starting 10848 1726776682.54906: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776681.1113408-10848-238830726360669/ > /dev/null 2>&1 && sleep 0' 10848 1726776682.57312: stderr chunk (state=2): >>><<< 10848 1726776682.57321: stdout chunk (state=2): >>><<< 10848 1726776682.57337: _low_level_execute_command() done: rc=0, stdout=, stderr= 10848 1726776682.57344: handler run complete 10848 1726776682.57363: Evaluated conditional (True): True 10848 1726776682.57374: attempt loop complete, returning result 10848 1726776682.57378: _execute() done 10848 1726776682.57383: dumping result to json 10848 1726776682.57388: done dumping result, returning 10848 1726776682.57397: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Tuned apply settings [120fa90a-8a95-fee7-0cea-0000000006ec] 10848 1726776682.57403: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ec 10848 1726776682.57435: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ec 10848 1726776682.57439: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "cmd": [ "tuned-adm", "profile", "virtual-guest kernel_settings" ], "delta": "0:00:01.170575", "end": "2024-09-19 16:11:22.533945", "rc": 0, "start": "2024-09-19 16:11:21.363370" } 9704 1726776682.57606: no more pending results, returning what we have 9704 1726776682.57610: results queue empty 9704 1726776682.57610: checking for any_errors_fatal 9704 1726776682.57618: done checking for any_errors_fatal 9704 1726776682.57618: checking for max_fail_percentage 9704 1726776682.57621: done checking for max_fail_percentage 9704 1726776682.57622: checking to see if all hosts have failed and the running result is not ok 9704 1726776682.57623: done checking to see if all hosts have failed 9704 1726776682.57624: getting the remaining hosts for this loop 9704 1726776682.57625: done getting the remaining hosts for this loop 9704 1726776682.57628: getting the next task for host managed_node3 9704 1726776682.57635: done getting next task for host managed_node3 9704 1726776682.57639: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Verify settings 9704 1726776682.57642: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=14, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776682.57652: getting variables 9704 1726776682.57654: in VariableManager get_vars() 9704 1726776682.57687: Calling all_inventory to load vars for managed_node3 9704 1726776682.57690: Calling groups_inventory to load vars for managed_node3 9704 1726776682.57691: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776682.57700: Calling all_plugins_play to load vars for managed_node3 9704 1726776682.57702: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776682.57704: Calling groups_plugins_play to load vars for managed_node3 9704 1726776682.57895: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776682.58015: done with get_vars() 9704 1726776682.58023: done getting variables TASK [fedora.linux_system_roles.kernel_settings : Verify settings] ************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:166 Thursday 19 September 2024 16:11:22 -0400 (0:00:01.550) 0:00:35.509 **** 9704 1726776682.58093: entering _queue_task() for managed_node3/include_tasks 9704 1726776682.58264: worker is 1 (out of 1 available) 9704 1726776682.58281: exiting _queue_task() for managed_node3/include_tasks 9704 1726776682.58293: done queuing things up, now waiting for results queue to drain 9704 1726776682.58295: waiting for pending results... 10892 1726776682.58426: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Verify settings 10892 1726776682.58555: in run() - task 120fa90a-8a95-fee7-0cea-0000000006ed 10892 1726776682.58572: variable 'ansible_search_path' from source: unknown 10892 1726776682.58576: variable 'ansible_search_path' from source: unknown 10892 1726776682.58604: calling self._execute() 10892 1726776682.58679: variable 'ansible_host' from source: host vars for 'managed_node3' 10892 1726776682.58688: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10892 1726776682.58697: variable 'omit' from source: magic vars 10892 1726776682.59024: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10892 1726776682.59209: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10892 1726776682.59246: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10892 1726776682.59275: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10892 1726776682.59302: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10892 1726776682.59388: variable '__kernel_settings_register_apply' from source: set_fact 10892 1726776682.59411: Evaluated conditional (__kernel_settings_register_apply is changed): True 10892 1726776682.59419: _execute() done 10892 1726776682.59424: dumping result to json 10892 1726776682.59427: done dumping result, returning 10892 1726776682.59436: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Verify settings [120fa90a-8a95-fee7-0cea-0000000006ed] 10892 1726776682.59442: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ed 10892 1726776682.59466: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ed 10892 1726776682.59470: WORKER PROCESS EXITING 9704 1726776682.59576: no more pending results, returning what we have 9704 1726776682.59581: in VariableManager get_vars() 9704 1726776682.59617: Calling all_inventory to load vars for managed_node3 9704 1726776682.59620: Calling groups_inventory to load vars for managed_node3 9704 1726776682.59622: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776682.59632: Calling all_plugins_play to load vars for managed_node3 9704 1726776682.59635: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776682.59637: Calling groups_plugins_play to load vars for managed_node3 9704 1726776682.59761: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776682.59880: done with get_vars() 9704 1726776682.59886: variable 'ansible_search_path' from source: unknown 9704 1726776682.59887: variable 'ansible_search_path' from source: unknown 9704 1726776682.59910: we have included files to process 9704 1726776682.59911: generating all_blocks data 9704 1726776682.59915: done generating all_blocks data 9704 1726776682.59919: processing included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml 9704 1726776682.59920: loading included file: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml 9704 1726776682.59921: Loading data from /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml included: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml for managed_node3 9704 1726776682.60192: done processing included file 9704 1726776682.60194: iterating over new_blocks loaded from include file 9704 1726776682.60194: in VariableManager get_vars() 9704 1726776682.60210: done with get_vars() 9704 1726776682.60211: filtering new block on tags 9704 1726776682.60251: done filtering new block on tags 9704 1726776682.60252: done iterating over new_blocks loaded from include file 9704 1726776682.60253: extending task lists for all hosts with included blocks 9704 1726776682.60827: done extending task lists 9704 1726776682.60829: done processing included files 9704 1726776682.60830: results queue empty 9704 1726776682.60830: checking for any_errors_fatal 9704 1726776682.60833: done checking for any_errors_fatal 9704 1726776682.60834: checking for max_fail_percentage 9704 1726776682.60834: done checking for max_fail_percentage 9704 1726776682.60835: checking to see if all hosts have failed and the running result is not ok 9704 1726776682.60835: done checking to see if all hosts have failed 9704 1726776682.60836: getting the remaining hosts for this loop 9704 1726776682.60836: done getting the remaining hosts for this loop 9704 1726776682.60838: getting the next task for host managed_node3 9704 1726776682.60841: done getting next task for host managed_node3 9704 1726776682.60843: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly 9704 1726776682.60845: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776682.60851: getting variables 9704 1726776682.60851: in VariableManager get_vars() 9704 1726776682.60861: Calling all_inventory to load vars for managed_node3 9704 1726776682.60862: Calling groups_inventory to load vars for managed_node3 9704 1726776682.60863: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776682.60867: Calling all_plugins_play to load vars for managed_node3 9704 1726776682.60869: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776682.60870: Calling groups_plugins_play to load vars for managed_node3 9704 1726776682.60946: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776682.61056: done with get_vars() 9704 1726776682.61063: done getting variables 9704 1726776682.61088: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:2 Thursday 19 September 2024 16:11:22 -0400 (0:00:00.030) 0:00:35.539 **** 9704 1726776682.61113: entering _queue_task() for managed_node3/command 9704 1726776682.61277: worker is 1 (out of 1 available) 9704 1726776682.61290: exiting _queue_task() for managed_node3/command 9704 1726776682.61301: done queuing things up, now waiting for results queue to drain 9704 1726776682.61303: waiting for pending results... 10893 1726776682.61433: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly 10893 1726776682.61594: in run() - task 120fa90a-8a95-fee7-0cea-0000000007cc 10893 1726776682.61612: variable 'ansible_search_path' from source: unknown 10893 1726776682.61616: variable 'ansible_search_path' from source: unknown 10893 1726776682.61644: calling self._execute() 10893 1726776682.61706: variable 'ansible_host' from source: host vars for 'managed_node3' 10893 1726776682.61712: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10893 1726776682.61718: variable 'omit' from source: magic vars 10893 1726776682.61790: variable 'omit' from source: magic vars 10893 1726776682.61838: variable 'omit' from source: magic vars 10893 1726776682.61862: variable 'omit' from source: magic vars 10893 1726776682.61891: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10893 1726776682.61915: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10893 1726776682.61934: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10893 1726776682.61949: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10893 1726776682.61960: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10893 1726776682.61980: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10893 1726776682.61984: variable 'ansible_host' from source: host vars for 'managed_node3' 10893 1726776682.61986: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10893 1726776682.62051: Set connection var ansible_module_compression to ZIP_DEFLATED 10893 1726776682.62060: Set connection var ansible_shell_type to sh 10893 1726776682.62065: Set connection var ansible_timeout to 10 10893 1726776682.62069: Set connection var ansible_pipelining to False 10893 1726776682.62073: Set connection var ansible_shell_executable to /bin/sh 10893 1726776682.62074: Set connection var ansible_connection to ssh 10893 1726776682.62088: variable 'ansible_shell_executable' from source: unknown 10893 1726776682.62091: variable 'ansible_connection' from source: unknown 10893 1726776682.62093: variable 'ansible_module_compression' from source: unknown 10893 1726776682.62095: variable 'ansible_shell_type' from source: unknown 10893 1726776682.62096: variable 'ansible_shell_executable' from source: unknown 10893 1726776682.62098: variable 'ansible_host' from source: host vars for 'managed_node3' 10893 1726776682.62100: variable 'ansible_pipelining' from source: unknown 10893 1726776682.62102: variable 'ansible_timeout' from source: unknown 10893 1726776682.62104: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10893 1726776682.62190: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10893 1726776682.62200: variable 'omit' from source: magic vars 10893 1726776682.62204: starting attempt loop 10893 1726776682.62206: running the handler 10893 1726776682.62216: _low_level_execute_command(): starting 10893 1726776682.62222: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10893 1726776682.64535: stdout chunk (state=2): >>>/root <<< 10893 1726776682.64654: stderr chunk (state=3): >>><<< 10893 1726776682.64663: stdout chunk (state=3): >>><<< 10893 1726776682.64680: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10893 1726776682.64693: _low_level_execute_command(): starting 10893 1726776682.64699: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309 `" && echo ansible-tmp-1726776682.646873-10893-258690486349309="` echo /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309 `" ) && sleep 0' 10893 1726776682.67107: stdout chunk (state=2): >>>ansible-tmp-1726776682.646873-10893-258690486349309=/root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309 <<< 10893 1726776682.67494: stderr chunk (state=3): >>><<< 10893 1726776682.67503: stdout chunk (state=3): >>><<< 10893 1726776682.67521: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776682.646873-10893-258690486349309=/root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309 , stderr= 10893 1726776682.67553: variable 'ansible_module_compression' from source: unknown 10893 1726776682.67615: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 10893 1726776682.67655: variable 'ansible_facts' from source: unknown 10893 1726776682.67765: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/AnsiballZ_command.py 10893 1726776682.68310: Sending initial data 10893 1726776682.68316: Sent initial data (154 bytes) 10893 1726776682.71067: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpcti21o8d /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/AnsiballZ_command.py <<< 10893 1726776682.72192: stderr chunk (state=3): >>><<< 10893 1726776682.72198: stdout chunk (state=3): >>><<< 10893 1726776682.72215: done transferring module to remote 10893 1726776682.72223: _low_level_execute_command(): starting 10893 1726776682.72226: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/ /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/AnsiballZ_command.py && sleep 0' 10893 1726776682.74687: stderr chunk (state=2): >>><<< 10893 1726776682.74695: stdout chunk (state=2): >>><<< 10893 1726776682.74711: _low_level_execute_command() done: rc=0, stdout=, stderr= 10893 1726776682.74716: _low_level_execute_command(): starting 10893 1726776682.74722: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/AnsiballZ_command.py && sleep 0' 10893 1726776683.00452: stdout chunk (state=2): >>> {"changed": true, "stdout": "Verification succeeded, current system settings match the preset profile.\nSee TuneD log file ('/var/log/tuned/tuned.log') for details.", "stderr": "", "rc": 0, "cmd": ["tuned-adm", "verify", "-i"], "start": "2024-09-19 16:11:22.897294", "end": "2024-09-19 16:11:23.002825", "delta": "0:00:00.105531", "msg": "", "invocation": {"module_args": {"_raw_params": "tuned-adm verify -i", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 10893 1726776683.01727: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10893 1726776683.01740: stdout chunk (state=3): >>><<< 10893 1726776683.01751: stderr chunk (state=3): >>><<< 10893 1726776683.01766: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "Verification succeeded, current system settings match the preset profile.\nSee TuneD log file ('/var/log/tuned/tuned.log') for details.", "stderr": "", "rc": 0, "cmd": ["tuned-adm", "verify", "-i"], "start": "2024-09-19 16:11:22.897294", "end": "2024-09-19 16:11:23.002825", "delta": "0:00:00.105531", "msg": "", "invocation": {"module_args": {"_raw_params": "tuned-adm verify -i", "_uses_shell": false, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10893 1726776683.01820: done with _execute_module (ansible.legacy.command, {'_raw_params': 'tuned-adm verify -i', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10893 1726776683.01834: _low_level_execute_command(): starting 10893 1726776683.01840: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776682.646873-10893-258690486349309/ > /dev/null 2>&1 && sleep 0' 10893 1726776683.04398: stderr chunk (state=2): >>><<< 10893 1726776683.04405: stdout chunk (state=2): >>><<< 10893 1726776683.04417: _low_level_execute_command() done: rc=0, stdout=, stderr= 10893 1726776683.04424: handler run complete 10893 1726776683.04443: Evaluated conditional (False): False 10893 1726776683.04453: attempt loop complete, returning result 10893 1726776683.04459: _execute() done 10893 1726776683.04462: dumping result to json 10893 1726776683.04468: done dumping result, returning 10893 1726776683.04475: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly [120fa90a-8a95-fee7-0cea-0000000007cc] 10893 1726776683.04483: sending task result for task 120fa90a-8a95-fee7-0cea-0000000007cc 10893 1726776683.04512: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000007cc 10893 1726776683.04516: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "cmd": [ "tuned-adm", "verify", "-i" ], "delta": "0:00:00.105531", "end": "2024-09-19 16:11:23.002825", "rc": 0, "start": "2024-09-19 16:11:22.897294" } STDOUT: Verification succeeded, current system settings match the preset profile. See TuneD log file ('/var/log/tuned/tuned.log') for details. 9704 1726776683.04765: no more pending results, returning what we have 9704 1726776683.04767: results queue empty 9704 1726776683.04768: checking for any_errors_fatal 9704 1726776683.04769: done checking for any_errors_fatal 9704 1726776683.04769: checking for max_fail_percentage 9704 1726776683.04770: done checking for max_fail_percentage 9704 1726776683.04771: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.04771: done checking to see if all hosts have failed 9704 1726776683.04772: getting the remaining hosts for this loop 9704 1726776683.04773: done getting the remaining hosts for this loop 9704 1726776683.04775: getting the next task for host managed_node3 9704 1726776683.04781: done getting next task for host managed_node3 9704 1726776683.04783: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Get last verify results from log 9704 1726776683.04786: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=2, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.04793: getting variables 9704 1726776683.04794: in VariableManager get_vars() 9704 1726776683.04820: Calling all_inventory to load vars for managed_node3 9704 1726776683.04822: Calling groups_inventory to load vars for managed_node3 9704 1726776683.04823: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.04831: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.04833: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.04834: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.04946: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.05116: done with get_vars() 9704 1726776683.05126: done getting variables 9704 1726776683.05192: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Get last verify results from log] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:12 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.441) 0:00:35.981 **** 9704 1726776683.05222: entering _queue_task() for managed_node3/shell 9704 1726776683.05381: worker is 1 (out of 1 available) 9704 1726776683.05394: exiting _queue_task() for managed_node3/shell 9704 1726776683.05405: done queuing things up, now waiting for results queue to drain 9704 1726776683.05408: waiting for pending results... 10915 1726776683.05557: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get last verify results from log 10915 1726776683.05743: in run() - task 120fa90a-8a95-fee7-0cea-0000000007cd 10915 1726776683.05761: variable 'ansible_search_path' from source: unknown 10915 1726776683.05765: variable 'ansible_search_path' from source: unknown 10915 1726776683.05795: calling self._execute() 10915 1726776683.05874: variable 'ansible_host' from source: host vars for 'managed_node3' 10915 1726776683.05883: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10915 1726776683.05891: variable 'omit' from source: magic vars 10915 1726776683.06289: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10915 1726776683.06468: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10915 1726776683.06505: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10915 1726776683.06533: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10915 1726776683.06560: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10915 1726776683.06638: variable '__kernel_settings_register_verify_values' from source: set_fact 10915 1726776683.06661: Evaluated conditional (__kernel_settings_register_verify_values is failed): False 10915 1726776683.06666: when evaluation is False, skipping this task 10915 1726776683.06669: _execute() done 10915 1726776683.06672: dumping result to json 10915 1726776683.06676: done dumping result, returning 10915 1726776683.06682: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Get last verify results from log [120fa90a-8a95-fee7-0cea-0000000007cd] 10915 1726776683.06688: sending task result for task 120fa90a-8a95-fee7-0cea-0000000007cd 10915 1726776683.06709: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000007cd 10915 1726776683.06713: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify_values is failed", "skip_reason": "Conditional result was False" } 9704 1726776683.06817: no more pending results, returning what we have 9704 1726776683.06820: results queue empty 9704 1726776683.06821: checking for any_errors_fatal 9704 1726776683.06831: done checking for any_errors_fatal 9704 1726776683.06831: checking for max_fail_percentage 9704 1726776683.06833: done checking for max_fail_percentage 9704 1726776683.06833: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.06834: done checking to see if all hosts have failed 9704 1726776683.06834: getting the remaining hosts for this loop 9704 1726776683.06836: done getting the remaining hosts for this loop 9704 1726776683.06839: getting the next task for host managed_node3 9704 1726776683.06846: done getting next task for host managed_node3 9704 1726776683.06849: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors 9704 1726776683.06853: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=15, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=3, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.06868: getting variables 9704 1726776683.06870: in VariableManager get_vars() 9704 1726776683.06897: Calling all_inventory to load vars for managed_node3 9704 1726776683.06899: Calling groups_inventory to load vars for managed_node3 9704 1726776683.06900: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.06907: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.06908: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.06910: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.07016: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.07140: done with get_vars() 9704 1726776683.07146: done getting variables 9704 1726776683.07187: Loading ActionModule 'fail' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/fail.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:23 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.019) 0:00:36.000 **** 9704 1726776683.07211: entering _queue_task() for managed_node3/fail 9704 1726776683.07366: worker is 1 (out of 1 available) 9704 1726776683.07379: exiting _queue_task() for managed_node3/fail 9704 1726776683.07389: done queuing things up, now waiting for results queue to drain 9704 1726776683.07391: waiting for pending results... 10917 1726776683.07509: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors 10917 1726776683.07634: in run() - task 120fa90a-8a95-fee7-0cea-0000000007ce 10917 1726776683.07650: variable 'ansible_search_path' from source: unknown 10917 1726776683.07654: variable 'ansible_search_path' from source: unknown 10917 1726776683.07680: calling self._execute() 10917 1726776683.07746: variable 'ansible_host' from source: host vars for 'managed_node3' 10917 1726776683.07754: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10917 1726776683.07763: variable 'omit' from source: magic vars 10917 1726776683.08080: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10917 1726776683.08315: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10917 1726776683.08349: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10917 1726776683.08375: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10917 1726776683.08402: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10917 1726776683.08480: variable '__kernel_settings_register_verify_values' from source: set_fact 10917 1726776683.08503: Evaluated conditional (__kernel_settings_register_verify_values is failed): False 10917 1726776683.08507: when evaluation is False, skipping this task 10917 1726776683.08509: _execute() done 10917 1726776683.08511: dumping result to json 10917 1726776683.08513: done dumping result, returning 10917 1726776683.08517: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Report errors that are not bootloader errors [120fa90a-8a95-fee7-0cea-0000000007ce] 10917 1726776683.08521: sending task result for task 120fa90a-8a95-fee7-0cea-0000000007ce 10917 1726776683.08541: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000007ce 10917 1726776683.08544: WORKER PROCESS EXITING skipping: [managed_node3] => { "changed": false, "false_condition": "__kernel_settings_register_verify_values is failed", "skip_reason": "Conditional result was False" } 9704 1726776683.08763: no more pending results, returning what we have 9704 1726776683.08765: results queue empty 9704 1726776683.08765: checking for any_errors_fatal 9704 1726776683.08769: done checking for any_errors_fatal 9704 1726776683.08769: checking for max_fail_percentage 9704 1726776683.08770: done checking for max_fail_percentage 9704 1726776683.08771: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.08771: done checking to see if all hosts have failed 9704 1726776683.08771: getting the remaining hosts for this loop 9704 1726776683.08772: done getting the remaining hosts for this loop 9704 1726776683.08774: getting the next task for host managed_node3 9704 1726776683.08780: done getting next task for host managed_node3 9704 1726776683.08783: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes 9704 1726776683.08788: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=16, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.08804: getting variables 9704 1726776683.08806: in VariableManager get_vars() 9704 1726776683.08837: Calling all_inventory to load vars for managed_node3 9704 1726776683.08839: Calling groups_inventory to load vars for managed_node3 9704 1726776683.08840: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.08847: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.08848: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.08850: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.09014: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.09133: done with get_vars() 9704 1726776683.09139: done getting variables 9704 1726776683.09181: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:177 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.019) 0:00:36.020 **** 9704 1726776683.09203: entering _queue_task() for managed_node3/set_fact 9704 1726776683.09383: worker is 1 (out of 1 available) 9704 1726776683.09395: exiting _queue_task() for managed_node3/set_fact 9704 1726776683.09406: done queuing things up, now waiting for results queue to drain 9704 1726776683.09408: waiting for pending results... 10918 1726776683.09627: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes 10918 1726776683.09784: in run() - task 120fa90a-8a95-fee7-0cea-0000000006ee 10918 1726776683.09800: variable 'ansible_search_path' from source: unknown 10918 1726776683.09804: variable 'ansible_search_path' from source: unknown 10918 1726776683.09839: calling self._execute() 10918 1726776683.09922: variable 'ansible_host' from source: host vars for 'managed_node3' 10918 1726776683.09932: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10918 1726776683.09941: variable 'omit' from source: magic vars 10918 1726776683.10039: variable 'omit' from source: magic vars 10918 1726776683.10090: variable 'omit' from source: magic vars 10918 1726776683.10122: variable 'omit' from source: magic vars 10918 1726776683.10165: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10918 1726776683.10199: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10918 1726776683.10223: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10918 1726776683.10244: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10918 1726776683.10259: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10918 1726776683.10288: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10918 1726776683.10294: variable 'ansible_host' from source: host vars for 'managed_node3' 10918 1726776683.10298: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10918 1726776683.10395: Set connection var ansible_module_compression to ZIP_DEFLATED 10918 1726776683.10404: Set connection var ansible_shell_type to sh 10918 1726776683.10411: Set connection var ansible_timeout to 10 10918 1726776683.10417: Set connection var ansible_pipelining to False 10918 1726776683.10424: Set connection var ansible_shell_executable to /bin/sh 10918 1726776683.10427: Set connection var ansible_connection to ssh 10918 1726776683.10474: variable 'ansible_shell_executable' from source: unknown 10918 1726776683.10479: variable 'ansible_connection' from source: unknown 10918 1726776683.10483: variable 'ansible_module_compression' from source: unknown 10918 1726776683.10486: variable 'ansible_shell_type' from source: unknown 10918 1726776683.10488: variable 'ansible_shell_executable' from source: unknown 10918 1726776683.10490: variable 'ansible_host' from source: host vars for 'managed_node3' 10918 1726776683.10494: variable 'ansible_pipelining' from source: unknown 10918 1726776683.10496: variable 'ansible_timeout' from source: unknown 10918 1726776683.10499: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10918 1726776683.10623: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10918 1726776683.10638: variable 'omit' from source: magic vars 10918 1726776683.10645: starting attempt loop 10918 1726776683.10648: running the handler 10918 1726776683.10660: handler run complete 10918 1726776683.10671: attempt loop complete, returning result 10918 1726776683.10675: _execute() done 10918 1726776683.10678: dumping result to json 10918 1726776683.10681: done dumping result, returning 10918 1726776683.10691: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set the flag that reboot is needed to apply changes [120fa90a-8a95-fee7-0cea-0000000006ee] 10918 1726776683.10698: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ee 10918 1726776683.10730: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ee 10918 1726776683.10734: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "kernel_settings_reboot_required": false }, "changed": false } 9704 1726776683.10930: no more pending results, returning what we have 9704 1726776683.10933: results queue empty 9704 1726776683.10933: checking for any_errors_fatal 9704 1726776683.10941: done checking for any_errors_fatal 9704 1726776683.10942: checking for max_fail_percentage 9704 1726776683.10943: done checking for max_fail_percentage 9704 1726776683.10944: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.10944: done checking to see if all hosts have failed 9704 1726776683.10945: getting the remaining hosts for this loop 9704 1726776683.10946: done getting the remaining hosts for this loop 9704 1726776683.10948: getting the next task for host managed_node3 9704 1726776683.10952: done getting next task for host managed_node3 9704 1726776683.10955: ^ task is: TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing 9704 1726776683.10957: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=4, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=17, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.10965: getting variables 9704 1726776683.10966: in VariableManager get_vars() 9704 1726776683.10989: Calling all_inventory to load vars for managed_node3 9704 1726776683.10991: Calling groups_inventory to load vars for managed_node3 9704 1726776683.10992: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.10998: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.11000: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.11001: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.11107: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.11231: done with get_vars() 9704 1726776683.11238: done getting variables 9704 1726776683.11276: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing] *** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:181 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.020) 0:00:36.041 **** 9704 1726776683.11298: entering _queue_task() for managed_node3/set_fact 9704 1726776683.11445: worker is 1 (out of 1 available) 9704 1726776683.11458: exiting _queue_task() for managed_node3/set_fact 9704 1726776683.11469: done queuing things up, now waiting for results queue to drain 9704 1726776683.11471: waiting for pending results... 10921 1726776683.11595: running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing 10921 1726776683.11710: in run() - task 120fa90a-8a95-fee7-0cea-0000000006ef 10921 1726776683.11726: variable 'ansible_search_path' from source: unknown 10921 1726776683.11732: variable 'ansible_search_path' from source: unknown 10921 1726776683.11760: calling self._execute() 10921 1726776683.11822: variable 'ansible_host' from source: host vars for 'managed_node3' 10921 1726776683.11831: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10921 1726776683.11840: variable 'omit' from source: magic vars 10921 1726776683.11913: variable 'omit' from source: magic vars 10921 1726776683.11952: variable 'omit' from source: magic vars 10921 1726776683.12206: '/usr/local/lib/python3.12/site-packages/ansible/plugins/test/__init__' skipped due to reserved name 10921 1726776683.12441: Loading TestModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/core.py 10921 1726776683.12478: Loading TestModule 'files' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/files.py 10921 1726776683.12502: Loading TestModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/mathstuff.py 10921 1726776683.12528: Loading TestModule 'uri' from /usr/local/lib/python3.12/site-packages/ansible/plugins/test/uri.py 10921 1726776683.12636: variable '__kernel_settings_register_profile' from source: set_fact 10921 1726776683.12649: variable '__kernel_settings_register_mode' from source: set_fact 10921 1726776683.12659: variable '__kernel_settings_register_apply' from source: set_fact 10921 1726776683.12705: variable 'omit' from source: magic vars 10921 1726776683.12732: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10921 1726776683.12763: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10921 1726776683.12780: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10921 1726776683.12793: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10921 1726776683.12805: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10921 1726776683.12840: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10921 1726776683.12846: variable 'ansible_host' from source: host vars for 'managed_node3' 10921 1726776683.12849: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10921 1726776683.12942: Set connection var ansible_module_compression to ZIP_DEFLATED 10921 1726776683.12952: Set connection var ansible_shell_type to sh 10921 1726776683.12962: Set connection var ansible_timeout to 10 10921 1726776683.12967: Set connection var ansible_pipelining to False 10921 1726776683.12977: Set connection var ansible_shell_executable to /bin/sh 10921 1726776683.12980: Set connection var ansible_connection to ssh 10921 1726776683.12998: variable 'ansible_shell_executable' from source: unknown 10921 1726776683.13001: variable 'ansible_connection' from source: unknown 10921 1726776683.13004: variable 'ansible_module_compression' from source: unknown 10921 1726776683.13007: variable 'ansible_shell_type' from source: unknown 10921 1726776683.13009: variable 'ansible_shell_executable' from source: unknown 10921 1726776683.13011: variable 'ansible_host' from source: host vars for 'managed_node3' 10921 1726776683.13015: variable 'ansible_pipelining' from source: unknown 10921 1726776683.13018: variable 'ansible_timeout' from source: unknown 10921 1726776683.13021: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10921 1726776683.13105: Loading ActionModule 'set_fact' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/set_fact.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10921 1726776683.13114: variable 'omit' from source: magic vars 10921 1726776683.13124: starting attempt loop 10921 1726776683.13128: running the handler 10921 1726776683.13141: handler run complete 10921 1726776683.13148: attempt loop complete, returning result 10921 1726776683.13151: _execute() done 10921 1726776683.13154: dumping result to json 10921 1726776683.13159: done dumping result, returning 10921 1726776683.13164: done running TaskExecutor() for managed_node3/TASK: fedora.linux_system_roles.kernel_settings : Set flag to indicate changed for testing [120fa90a-8a95-fee7-0cea-0000000006ef] 10921 1726776683.13169: sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ef 10921 1726776683.13190: done sending task result for task 120fa90a-8a95-fee7-0cea-0000000006ef 10921 1726776683.13193: WORKER PROCESS EXITING ok: [managed_node3] => { "ansible_facts": { "__kernel_settings_changed": true }, "changed": false } 9704 1726776683.13563: no more pending results, returning what we have 9704 1726776683.13566: results queue empty 9704 1726776683.13567: checking for any_errors_fatal 9704 1726776683.13576: done checking for any_errors_fatal 9704 1726776683.13577: checking for max_fail_percentage 9704 1726776683.13578: done checking for max_fail_percentage 9704 1726776683.13579: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.13580: done checking to see if all hosts have failed 9704 1726776683.13580: getting the remaining hosts for this loop 9704 1726776683.13582: done getting the remaining hosts for this loop 9704 1726776683.13584: getting the next task for host managed_node3 9704 1726776683.13593: done getting next task for host managed_node3 9704 1726776683.13594: ^ task is: TASK: meta (role_complete) 9704 1726776683.13597: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=5, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (HOST STATE: block=0, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.13610: getting variables 9704 1726776683.13613: in VariableManager get_vars() 9704 1726776683.13646: Calling all_inventory to load vars for managed_node3 9704 1726776683.13649: Calling groups_inventory to load vars for managed_node3 9704 1726776683.13651: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.13661: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.13663: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.13666: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.13813: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.14038: done with get_vars() 9704 1726776683.14046: done getting variables 9704 1726776683.14121: done queuing things up, now waiting for results queue to drain 9704 1726776683.14127: results queue empty 9704 1726776683.14128: checking for any_errors_fatal 9704 1726776683.14133: done checking for any_errors_fatal 9704 1726776683.14134: checking for max_fail_percentage 9704 1726776683.14135: done checking for max_fail_percentage 9704 1726776683.14135: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.14136: done checking to see if all hosts have failed 9704 1726776683.14136: getting the remaining hosts for this loop 9704 1726776683.14137: done getting the remaining hosts for this loop 9704 1726776683.14139: getting the next task for host managed_node3 9704 1726776683.14142: done getting next task for host managed_node3 9704 1726776683.14144: ^ task is: TASK: Verify no settings 9704 1726776683.14145: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.14147: getting variables 9704 1726776683.14148: in VariableManager get_vars() 9704 1726776683.14161: Calling all_inventory to load vars for managed_node3 9704 1726776683.14163: Calling groups_inventory to load vars for managed_node3 9704 1726776683.14165: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.14170: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.14172: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.14173: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.14262: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.14449: done with get_vars() 9704 1726776683.14460: done getting variables 9704 1726776683.14494: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Verify no settings] ****************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:20 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.032) 0:00:36.073 **** 9704 1726776683.14519: entering _queue_task() for managed_node3/shell 9704 1726776683.14736: worker is 1 (out of 1 available) 9704 1726776683.14751: exiting _queue_task() for managed_node3/shell 9704 1726776683.14766: done queuing things up, now waiting for results queue to drain 9704 1726776683.14769: waiting for pending results... 10923 1726776683.14926: running TaskExecutor() for managed_node3/TASK: Verify no settings 10923 1726776683.15033: in run() - task 120fa90a-8a95-fee7-0cea-00000000058b 10923 1726776683.15050: variable 'ansible_search_path' from source: unknown 10923 1726776683.15056: variable 'ansible_search_path' from source: unknown 10923 1726776683.15084: calling self._execute() 10923 1726776683.15155: variable 'ansible_host' from source: host vars for 'managed_node3' 10923 1726776683.15164: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10923 1726776683.15173: variable 'omit' from source: magic vars 10923 1726776683.15249: variable 'omit' from source: magic vars 10923 1726776683.15281: variable 'omit' from source: magic vars 10923 1726776683.15522: variable '__kernel_settings_profile_filename' from source: role '' exported vars 10923 1726776683.15582: variable '__kernel_settings_profile_dir' from source: role '' exported vars 10923 1726776683.15645: variable '__kernel_settings_profile_parent' from source: set_fact 10923 1726776683.15654: variable '__kernel_settings_tuned_profile' from source: role '' exported vars 10923 1726776683.15689: variable 'omit' from source: magic vars 10923 1726776683.15716: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10923 1726776683.15743: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10923 1726776683.15761: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10923 1726776683.15773: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10923 1726776683.15781: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10923 1726776683.15802: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10923 1726776683.15806: variable 'ansible_host' from source: host vars for 'managed_node3' 10923 1726776683.15808: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10923 1726776683.15890: Set connection var ansible_module_compression to ZIP_DEFLATED 10923 1726776683.15899: Set connection var ansible_shell_type to sh 10923 1726776683.15907: Set connection var ansible_timeout to 10 10923 1726776683.15912: Set connection var ansible_pipelining to False 10923 1726776683.15919: Set connection var ansible_shell_executable to /bin/sh 10923 1726776683.15922: Set connection var ansible_connection to ssh 10923 1726776683.15940: variable 'ansible_shell_executable' from source: unknown 10923 1726776683.15944: variable 'ansible_connection' from source: unknown 10923 1726776683.15947: variable 'ansible_module_compression' from source: unknown 10923 1726776683.15950: variable 'ansible_shell_type' from source: unknown 10923 1726776683.15954: variable 'ansible_shell_executable' from source: unknown 10923 1726776683.15957: variable 'ansible_host' from source: host vars for 'managed_node3' 10923 1726776683.15962: variable 'ansible_pipelining' from source: unknown 10923 1726776683.15965: variable 'ansible_timeout' from source: unknown 10923 1726776683.15969: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10923 1726776683.16059: Loading ActionModule 'shell' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/shell.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10923 1726776683.16071: variable 'omit' from source: magic vars 10923 1726776683.16077: starting attempt loop 10923 1726776683.16081: running the handler 10923 1726776683.16088: Loading ActionModule 'command' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/command.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10923 1726776683.16101: _low_level_execute_command(): starting 10923 1726776683.16106: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10923 1726776683.18550: stdout chunk (state=2): >>>/root <<< 10923 1726776683.18665: stderr chunk (state=3): >>><<< 10923 1726776683.18672: stdout chunk (state=3): >>><<< 10923 1726776683.18689: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10923 1726776683.18704: _low_level_execute_command(): starting 10923 1726776683.18710: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590 `" && echo ansible-tmp-1726776683.1869714-10923-17929259252590="` echo /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590 `" ) && sleep 0' 10923 1726776683.21082: stdout chunk (state=2): >>>ansible-tmp-1726776683.1869714-10923-17929259252590=/root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590 <<< 10923 1726776683.21211: stderr chunk (state=3): >>><<< 10923 1726776683.21218: stdout chunk (state=3): >>><<< 10923 1726776683.21235: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776683.1869714-10923-17929259252590=/root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590 , stderr= 10923 1726776683.21262: variable 'ansible_module_compression' from source: unknown 10923 1726776683.21302: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.command-ZIP_DEFLATED 10923 1726776683.21325: variable 'ansible_facts' from source: unknown 10923 1726776683.21397: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/AnsiballZ_command.py 10923 1726776683.21558: Sending initial data 10923 1726776683.21565: Sent initial data (154 bytes) 10923 1726776683.24043: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp7tspj_aa /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/AnsiballZ_command.py <<< 10923 1726776683.25020: stderr chunk (state=3): >>><<< 10923 1726776683.25027: stdout chunk (state=3): >>><<< 10923 1726776683.25048: done transferring module to remote 10923 1726776683.25060: _low_level_execute_command(): starting 10923 1726776683.25066: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/ /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/AnsiballZ_command.py && sleep 0' 10923 1726776683.27410: stderr chunk (state=2): >>><<< 10923 1726776683.27419: stdout chunk (state=2): >>><<< 10923 1726776683.27435: _low_level_execute_command() done: rc=0, stdout=, stderr= 10923 1726776683.27439: _low_level_execute_command(): starting 10923 1726776683.27444: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/AnsiballZ_command.py && sleep 0' 10923 1726776683.43335: stdout chunk (state=2): >>> {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ conf=/etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[sysctl\\]' /etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[sysfs\\]' /etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[systemd\\]' /etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[vm\\]' /etc/tuned/kernel_settings/tuned.conf\n+ exit 0", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nconf=/etc/tuned/kernel_settings/tuned.conf\nfor section in sysctl sysfs systemd vm; do\n if grep ^\\\\[\"$section\"\\\\] \"$conf\"; then\n echo ERROR: \"$section\" settings present\n rc=1\n fi\ndone\nexit \"$rc\"\n", "start": "2024-09-19 16:11:23.424450", "end": "2024-09-19 16:11:23.431838", "delta": "0:00:00.007388", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nconf=/etc/tuned/kernel_settings/tuned.conf\nfor section in sysctl sysfs systemd vm; do\n if grep ^\\\\[\"$section\"\\\\] \"$conf\"; then\n echo ERROR: \"$section\" settings present\n rc=1\n fi\ndone\nexit \"$rc\"\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} <<< 10923 1726776683.44500: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10923 1726776683.44557: stderr chunk (state=3): >>><<< 10923 1726776683.44566: stdout chunk (state=3): >>><<< 10923 1726776683.44582: _low_level_execute_command() done: rc=0, stdout= {"changed": true, "stdout": "", "stderr": "+ exec\n+ rc=0\n+ conf=/etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[sysctl\\]' /etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[sysfs\\]' /etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[systemd\\]' /etc/tuned/kernel_settings/tuned.conf\n+ for section in sysctl sysfs systemd vm\n+ grep '^\\[vm\\]' /etc/tuned/kernel_settings/tuned.conf\n+ exit 0", "rc": 0, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nconf=/etc/tuned/kernel_settings/tuned.conf\nfor section in sysctl sysfs systemd vm; do\n if grep ^\\\\[\"$section\"\\\\] \"$conf\"; then\n echo ERROR: \"$section\" settings present\n rc=1\n fi\ndone\nexit \"$rc\"\n", "start": "2024-09-19 16:11:23.424450", "end": "2024-09-19 16:11:23.431838", "delta": "0:00:00.007388", "msg": "", "invocation": {"module_args": {"_raw_params": "set -euxo pipefail\nexec 1>&2\nrc=0\nconf=/etc/tuned/kernel_settings/tuned.conf\nfor section in sysctl sysfs systemd vm; do\n if grep ^\\\\[\"$section\"\\\\] \"$conf\"; then\n echo ERROR: \"$section\" settings present\n rc=1\n fi\ndone\nexit \"$rc\"\n", "_uses_shell": true, "expand_argument_vars": true, "stdin_add_newline": true, "strip_empty_ends": true, "argv": null, "chdir": null, "executable": null, "creates": null, "removes": null, "stdin": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10923 1726776683.44612: done with _execute_module (ansible.legacy.command, {'_raw_params': 'set -euxo pipefail\nexec 1>&2\nrc=0\nconf=/etc/tuned/kernel_settings/tuned.conf\nfor section in sysctl sysfs systemd vm; do\n if grep ^\\\\["$section"\\\\] "$conf"; then\n echo ERROR: "$section" settings present\n rc=1\n fi\ndone\nexit "$rc"\n', '_uses_shell': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.command', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10923 1726776683.44622: _low_level_execute_command(): starting 10923 1726776683.44630: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776683.1869714-10923-17929259252590/ > /dev/null 2>&1 && sleep 0' 10923 1726776683.46991: stderr chunk (state=2): >>><<< 10923 1726776683.46998: stdout chunk (state=2): >>><<< 10923 1726776683.47011: _low_level_execute_command() done: rc=0, stdout=, stderr= 10923 1726776683.47026: handler run complete 10923 1726776683.47058: Evaluated conditional (False): False 10923 1726776683.47070: attempt loop complete, returning result 10923 1726776683.47076: _execute() done 10923 1726776683.47079: dumping result to json 10923 1726776683.47084: done dumping result, returning 10923 1726776683.47090: done running TaskExecutor() for managed_node3/TASK: Verify no settings [120fa90a-8a95-fee7-0cea-00000000058b] 10923 1726776683.47097: sending task result for task 120fa90a-8a95-fee7-0cea-00000000058b 10923 1726776683.47140: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000058b 10923 1726776683.47144: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "cmd": "set -euxo pipefail\nexec 1>&2\nrc=0\nconf=/etc/tuned/kernel_settings/tuned.conf\nfor section in sysctl sysfs systemd vm; do\n if grep ^\\\\[\"$section\"\\\\] \"$conf\"; then\n echo ERROR: \"$section\" settings present\n rc=1\n fi\ndone\nexit \"$rc\"\n", "delta": "0:00:00.007388", "end": "2024-09-19 16:11:23.431838", "rc": 0, "start": "2024-09-19 16:11:23.424450" } STDERR: + exec + rc=0 + conf=/etc/tuned/kernel_settings/tuned.conf + for section in sysctl sysfs systemd vm + grep '^\[sysctl\]' /etc/tuned/kernel_settings/tuned.conf + for section in sysctl sysfs systemd vm + grep '^\[sysfs\]' /etc/tuned/kernel_settings/tuned.conf + for section in sysctl sysfs systemd vm + grep '^\[systemd\]' /etc/tuned/kernel_settings/tuned.conf + for section in sysctl sysfs systemd vm + grep '^\[vm\]' /etc/tuned/kernel_settings/tuned.conf + exit 0 9704 1726776683.47696: no more pending results, returning what we have 9704 1726776683.47701: results queue empty 9704 1726776683.47701: checking for any_errors_fatal 9704 1726776683.47703: done checking for any_errors_fatal 9704 1726776683.47704: checking for max_fail_percentage 9704 1726776683.47705: done checking for max_fail_percentage 9704 1726776683.47706: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.47707: done checking to see if all hosts have failed 9704 1726776683.47707: getting the remaining hosts for this loop 9704 1726776683.47709: done getting the remaining hosts for this loop 9704 1726776683.47712: getting the next task for host managed_node3 9704 1726776683.47719: done getting next task for host managed_node3 9704 1726776683.47721: ^ task is: TASK: Remove kernel_settings tuned profile 9704 1726776683.47723: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=1, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.47727: getting variables 9704 1726776683.47730: in VariableManager get_vars() 9704 1726776683.47767: Calling all_inventory to load vars for managed_node3 9704 1726776683.47771: Calling groups_inventory to load vars for managed_node3 9704 1726776683.47773: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.47783: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.47790: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.47793: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.47973: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.48230: done with get_vars() 9704 1726776683.48240: done getting variables TASK [Remove kernel_settings tuned profile] ************************************ task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:36 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.338) 0:00:36.412 **** 9704 1726776683.48323: entering _queue_task() for managed_node3/file 9704 1726776683.48493: worker is 1 (out of 1 available) 9704 1726776683.48507: exiting _queue_task() for managed_node3/file 9704 1726776683.48517: done queuing things up, now waiting for results queue to drain 9704 1726776683.48519: waiting for pending results... 10939 1726776683.48651: running TaskExecutor() for managed_node3/TASK: Remove kernel_settings tuned profile 10939 1726776683.48760: in run() - task 120fa90a-8a95-fee7-0cea-00000000058c 10939 1726776683.48776: variable 'ansible_search_path' from source: unknown 10939 1726776683.48780: variable 'ansible_search_path' from source: unknown 10939 1726776683.48807: calling self._execute() 10939 1726776683.48879: variable 'ansible_host' from source: host vars for 'managed_node3' 10939 1726776683.48887: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10939 1726776683.48896: variable 'omit' from source: magic vars 10939 1726776683.48973: variable 'omit' from source: magic vars 10939 1726776683.49001: variable 'omit' from source: magic vars 10939 1726776683.49024: variable '__kernel_settings_profile_dir' from source: role '' exported vars 10939 1726776683.49237: variable '__kernel_settings_profile_dir' from source: role '' exported vars 10939 1726776683.49306: variable '__kernel_settings_profile_parent' from source: set_fact 10939 1726776683.49315: variable '__kernel_settings_tuned_profile' from source: role '' exported vars 10939 1726776683.49348: variable 'omit' from source: magic vars 10939 1726776683.49379: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10939 1726776683.49407: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10939 1726776683.49427: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10939 1726776683.49442: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10939 1726776683.49452: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10939 1726776683.49475: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10939 1726776683.49480: variable 'ansible_host' from source: host vars for 'managed_node3' 10939 1726776683.49484: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10939 1726776683.49553: Set connection var ansible_module_compression to ZIP_DEFLATED 10939 1726776683.49563: Set connection var ansible_shell_type to sh 10939 1726776683.49572: Set connection var ansible_timeout to 10 10939 1726776683.49577: Set connection var ansible_pipelining to False 10939 1726776683.49584: Set connection var ansible_shell_executable to /bin/sh 10939 1726776683.49587: Set connection var ansible_connection to ssh 10939 1726776683.49602: variable 'ansible_shell_executable' from source: unknown 10939 1726776683.49606: variable 'ansible_connection' from source: unknown 10939 1726776683.49611: variable 'ansible_module_compression' from source: unknown 10939 1726776683.49614: variable 'ansible_shell_type' from source: unknown 10939 1726776683.49618: variable 'ansible_shell_executable' from source: unknown 10939 1726776683.49621: variable 'ansible_host' from source: host vars for 'managed_node3' 10939 1726776683.49625: variable 'ansible_pipelining' from source: unknown 10939 1726776683.49629: variable 'ansible_timeout' from source: unknown 10939 1726776683.49634: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10939 1726776683.49769: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10939 1726776683.49780: variable 'omit' from source: magic vars 10939 1726776683.49786: starting attempt loop 10939 1726776683.49790: running the handler 10939 1726776683.49801: _low_level_execute_command(): starting 10939 1726776683.49809: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10939 1726776683.52081: stdout chunk (state=2): >>>/root <<< 10939 1726776683.52197: stderr chunk (state=3): >>><<< 10939 1726776683.52204: stdout chunk (state=3): >>><<< 10939 1726776683.52223: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10939 1726776683.52240: _low_level_execute_command(): starting 10939 1726776683.52247: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062 `" && echo ansible-tmp-1726776683.5223262-10939-229357604384062="` echo /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062 `" ) && sleep 0' 10939 1726776683.54649: stdout chunk (state=2): >>>ansible-tmp-1726776683.5223262-10939-229357604384062=/root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062 <<< 10939 1726776683.54766: stderr chunk (state=3): >>><<< 10939 1726776683.54773: stdout chunk (state=3): >>><<< 10939 1726776683.54787: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776683.5223262-10939-229357604384062=/root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062 , stderr= 10939 1726776683.54822: variable 'ansible_module_compression' from source: unknown 10939 1726776683.54867: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.file-ZIP_DEFLATED 10939 1726776683.54895: variable 'ansible_facts' from source: unknown 10939 1726776683.54966: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/AnsiballZ_file.py 10939 1726776683.55065: Sending initial data 10939 1726776683.55073: Sent initial data (152 bytes) 10939 1726776683.57475: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpgx0qdgav /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/AnsiballZ_file.py <<< 10939 1726776683.58473: stderr chunk (state=3): >>><<< 10939 1726776683.58480: stdout chunk (state=3): >>><<< 10939 1726776683.58498: done transferring module to remote 10939 1726776683.58509: _low_level_execute_command(): starting 10939 1726776683.58514: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/ /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/AnsiballZ_file.py && sleep 0' 10939 1726776683.60881: stderr chunk (state=2): >>><<< 10939 1726776683.60888: stdout chunk (state=2): >>><<< 10939 1726776683.60901: _low_level_execute_command() done: rc=0, stdout=, stderr= 10939 1726776683.60905: _low_level_execute_command(): starting 10939 1726776683.60909: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/AnsiballZ_file.py && sleep 0' 10939 1726776683.76612: stdout chunk (state=2): >>> {"path": "/etc/tuned/kernel_settings", "changed": true, "diff": {"before": {"path": "/etc/tuned/kernel_settings", "state": "directory", "path_content": {"directories": [], "files": ["/etc/tuned/kernel_settings/tuned.conf"]}}, "after": {"path": "/etc/tuned/kernel_settings", "state": "absent"}}, "state": "absent", "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "absent", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "mode": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10939 1726776683.77650: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10939 1726776683.77690: stderr chunk (state=3): >>><<< 10939 1726776683.77698: stdout chunk (state=3): >>><<< 10939 1726776683.77716: _low_level_execute_command() done: rc=0, stdout= {"path": "/etc/tuned/kernel_settings", "changed": true, "diff": {"before": {"path": "/etc/tuned/kernel_settings", "state": "directory", "path_content": {"directories": [], "files": ["/etc/tuned/kernel_settings/tuned.conf"]}}, "after": {"path": "/etc/tuned/kernel_settings", "state": "absent"}}, "state": "absent", "invocation": {"module_args": {"path": "/etc/tuned/kernel_settings", "state": "absent", "recurse": false, "force": false, "follow": true, "modification_time_format": "%Y%m%d%H%M.%S", "access_time_format": "%Y%m%d%H%M.%S", "unsafe_writes": false, "_original_basename": null, "_diff_peek": null, "src": null, "modification_time": null, "access_time": null, "mode": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10939 1726776683.77750: done with _execute_module (file, {'path': '/etc/tuned/kernel_settings', 'state': 'absent', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'file', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10939 1726776683.77762: _low_level_execute_command(): starting 10939 1726776683.77769: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776683.5223262-10939-229357604384062/ > /dev/null 2>&1 && sleep 0' 10939 1726776683.80133: stderr chunk (state=2): >>><<< 10939 1726776683.80141: stdout chunk (state=2): >>><<< 10939 1726776683.80155: _low_level_execute_command() done: rc=0, stdout=, stderr= 10939 1726776683.80165: handler run complete 10939 1726776683.80186: attempt loop complete, returning result 10939 1726776683.80190: _execute() done 10939 1726776683.80193: dumping result to json 10939 1726776683.80199: done dumping result, returning 10939 1726776683.80206: done running TaskExecutor() for managed_node3/TASK: Remove kernel_settings tuned profile [120fa90a-8a95-fee7-0cea-00000000058c] 10939 1726776683.80214: sending task result for task 120fa90a-8a95-fee7-0cea-00000000058c 10939 1726776683.80248: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000058c 10939 1726776683.80252: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "path": "/etc/tuned/kernel_settings", "state": "absent" } 9704 1726776683.80388: no more pending results, returning what we have 9704 1726776683.80392: results queue empty 9704 1726776683.80392: checking for any_errors_fatal 9704 1726776683.80400: done checking for any_errors_fatal 9704 1726776683.80401: checking for max_fail_percentage 9704 1726776683.80402: done checking for max_fail_percentage 9704 1726776683.80403: checking to see if all hosts have failed and the running result is not ok 9704 1726776683.80404: done checking to see if all hosts have failed 9704 1726776683.80404: getting the remaining hosts for this loop 9704 1726776683.80405: done getting the remaining hosts for this loop 9704 1726776683.80408: getting the next task for host managed_node3 9704 1726776683.80414: done getting next task for host managed_node3 9704 1726776683.80416: ^ task is: TASK: Get active_profile 9704 1726776683.80419: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=2, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776683.80423: getting variables 9704 1726776683.80424: in VariableManager get_vars() 9704 1726776683.80459: Calling all_inventory to load vars for managed_node3 9704 1726776683.80462: Calling groups_inventory to load vars for managed_node3 9704 1726776683.80463: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776683.80472: Calling all_plugins_play to load vars for managed_node3 9704 1726776683.80475: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776683.80477: Calling groups_plugins_play to load vars for managed_node3 9704 1726776683.80599: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776683.80717: done with get_vars() 9704 1726776683.80725: done getting variables TASK [Get active_profile] ****************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:41 Thursday 19 September 2024 16:11:23 -0400 (0:00:00.324) 0:00:36.736 **** 9704 1726776683.80792: entering _queue_task() for managed_node3/slurp 9704 1726776683.80958: worker is 1 (out of 1 available) 9704 1726776683.80972: exiting _queue_task() for managed_node3/slurp 9704 1726776683.80983: done queuing things up, now waiting for results queue to drain 9704 1726776683.80985: waiting for pending results... 10953 1726776683.81116: running TaskExecutor() for managed_node3/TASK: Get active_profile 10953 1726776683.81231: in run() - task 120fa90a-8a95-fee7-0cea-00000000058d 10953 1726776683.81248: variable 'ansible_search_path' from source: unknown 10953 1726776683.81252: variable 'ansible_search_path' from source: unknown 10953 1726776683.81281: calling self._execute() 10953 1726776683.81351: variable 'ansible_host' from source: host vars for 'managed_node3' 10953 1726776683.81361: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10953 1726776683.81371: variable 'omit' from source: magic vars 10953 1726776683.81448: variable 'omit' from source: magic vars 10953 1726776683.81483: variable 'omit' from source: magic vars 10953 1726776683.81505: variable '__kernel_settings_tuned_active_profile' from source: role '' exported vars 10953 1726776683.81718: variable '__kernel_settings_tuned_active_profile' from source: role '' exported vars 10953 1726776683.81782: variable '__kernel_settings_tuned_dir' from source: role '' exported vars 10953 1726776683.81811: variable 'omit' from source: magic vars 10953 1726776683.81844: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10953 1726776683.81873: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10953 1726776683.81891: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10953 1726776683.81963: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10953 1726776683.81975: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10953 1726776683.81998: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10953 1726776683.82002: variable 'ansible_host' from source: host vars for 'managed_node3' 10953 1726776683.82007: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10953 1726776683.82076: Set connection var ansible_module_compression to ZIP_DEFLATED 10953 1726776683.82084: Set connection var ansible_shell_type to sh 10953 1726776683.82092: Set connection var ansible_timeout to 10 10953 1726776683.82098: Set connection var ansible_pipelining to False 10953 1726776683.82105: Set connection var ansible_shell_executable to /bin/sh 10953 1726776683.82108: Set connection var ansible_connection to ssh 10953 1726776683.82124: variable 'ansible_shell_executable' from source: unknown 10953 1726776683.82128: variable 'ansible_connection' from source: unknown 10953 1726776683.82132: variable 'ansible_module_compression' from source: unknown 10953 1726776683.82136: variable 'ansible_shell_type' from source: unknown 10953 1726776683.82140: variable 'ansible_shell_executable' from source: unknown 10953 1726776683.82143: variable 'ansible_host' from source: host vars for 'managed_node3' 10953 1726776683.82147: variable 'ansible_pipelining' from source: unknown 10953 1726776683.82150: variable 'ansible_timeout' from source: unknown 10953 1726776683.82155: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10953 1726776683.82293: Loading ActionModule 'normal' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/normal.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) 10953 1726776683.82303: variable 'omit' from source: magic vars 10953 1726776683.82309: starting attempt loop 10953 1726776683.82312: running the handler 10953 1726776683.82323: _low_level_execute_command(): starting 10953 1726776683.82332: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10953 1726776683.84822: stdout chunk (state=2): >>>/root <<< 10953 1726776683.84944: stderr chunk (state=3): >>><<< 10953 1726776683.84952: stdout chunk (state=3): >>><<< 10953 1726776683.84972: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10953 1726776683.84987: _low_level_execute_command(): starting 10953 1726776683.84994: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161 `" && echo ansible-tmp-1726776683.8498106-10953-221251324612161="` echo /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161 `" ) && sleep 0' 10953 1726776683.87442: stdout chunk (state=2): >>>ansible-tmp-1726776683.8498106-10953-221251324612161=/root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161 <<< 10953 1726776683.87571: stderr chunk (state=3): >>><<< 10953 1726776683.87581: stdout chunk (state=3): >>><<< 10953 1726776683.87597: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776683.8498106-10953-221251324612161=/root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161 , stderr= 10953 1726776683.87636: variable 'ansible_module_compression' from source: unknown 10953 1726776683.87670: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.slurp-ZIP_DEFLATED 10953 1726776683.87698: variable 'ansible_facts' from source: unknown 10953 1726776683.87769: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/AnsiballZ_slurp.py 10953 1726776683.87874: Sending initial data 10953 1726776683.87881: Sent initial data (153 bytes) 10953 1726776683.90339: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpmkrbnmi5 /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/AnsiballZ_slurp.py <<< 10953 1726776683.91280: stderr chunk (state=3): >>><<< 10953 1726776683.91287: stdout chunk (state=3): >>><<< 10953 1726776683.91303: done transferring module to remote 10953 1726776683.91314: _low_level_execute_command(): starting 10953 1726776683.91319: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/ /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/AnsiballZ_slurp.py && sleep 0' 10953 1726776683.93667: stderr chunk (state=2): >>><<< 10953 1726776683.93673: stdout chunk (state=2): >>><<< 10953 1726776683.93685: _low_level_execute_command() done: rc=0, stdout=, stderr= 10953 1726776683.93689: _low_level_execute_command(): starting 10953 1726776683.93696: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/AnsiballZ_slurp.py && sleep 0' 10953 1726776684.08454: stdout chunk (state=2): >>> {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} <<< 10953 1726776684.09556: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10953 1726776684.09606: stderr chunk (state=3): >>><<< 10953 1726776684.09613: stdout chunk (state=3): >>><<< 10953 1726776684.09634: _low_level_execute_command() done: rc=0, stdout= {"content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "source": "/etc/tuned/active_profile", "encoding": "base64", "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "src": "/etc/tuned/active_profile"}}} , stderr=Shared connection to 10.31.8.186 closed. 10953 1726776684.09660: done with _execute_module (slurp, {'path': '/etc/tuned/active_profile', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'slurp', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10953 1726776684.09674: _low_level_execute_command(): starting 10953 1726776684.09682: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776683.8498106-10953-221251324612161/ > /dev/null 2>&1 && sleep 0' 10953 1726776684.12150: stderr chunk (state=2): >>><<< 10953 1726776684.12161: stdout chunk (state=2): >>><<< 10953 1726776684.12175: _low_level_execute_command() done: rc=0, stdout=, stderr= 10953 1726776684.12183: handler run complete 10953 1726776684.12196: attempt loop complete, returning result 10953 1726776684.12200: _execute() done 10953 1726776684.12203: dumping result to json 10953 1726776684.12207: done dumping result, returning 10953 1726776684.12215: done running TaskExecutor() for managed_node3/TASK: Get active_profile [120fa90a-8a95-fee7-0cea-00000000058d] 10953 1726776684.12221: sending task result for task 120fa90a-8a95-fee7-0cea-00000000058d 10953 1726776684.12252: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000058d 10953 1726776684.12256: WORKER PROCESS EXITING ok: [managed_node3] => { "changed": false, "content": "dmlydHVhbC1ndWVzdCBrZXJuZWxfc2V0dGluZ3MK", "encoding": "base64", "source": "/etc/tuned/active_profile" } 9704 1726776684.12389: no more pending results, returning what we have 9704 1726776684.12392: results queue empty 9704 1726776684.12393: checking for any_errors_fatal 9704 1726776684.12401: done checking for any_errors_fatal 9704 1726776684.12402: checking for max_fail_percentage 9704 1726776684.12403: done checking for max_fail_percentage 9704 1726776684.12404: checking to see if all hosts have failed and the running result is not ok 9704 1726776684.12404: done checking to see if all hosts have failed 9704 1726776684.12405: getting the remaining hosts for this loop 9704 1726776684.12406: done getting the remaining hosts for this loop 9704 1726776684.12409: getting the next task for host managed_node3 9704 1726776684.12416: done getting next task for host managed_node3 9704 1726776684.12418: ^ task is: TASK: Ensure kernel_settings is not in active_profile 9704 1726776684.12420: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776684.12426: getting variables 9704 1726776684.12430: in VariableManager get_vars() 9704 1726776684.12463: Calling all_inventory to load vars for managed_node3 9704 1726776684.12466: Calling groups_inventory to load vars for managed_node3 9704 1726776684.12468: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776684.12477: Calling all_plugins_play to load vars for managed_node3 9704 1726776684.12480: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776684.12482: Calling groups_plugins_play to load vars for managed_node3 9704 1726776684.12649: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776684.12904: done with get_vars() 9704 1726776684.12913: done getting variables 9704 1726776684.12973: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Ensure kernel_settings is not in active_profile] ************************* task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:46 Thursday 19 September 2024 16:11:24 -0400 (0:00:00.322) 0:00:37.058 **** 9704 1726776684.12999: entering _queue_task() for managed_node3/copy 9704 1726776684.13208: worker is 1 (out of 1 available) 9704 1726776684.13225: exiting _queue_task() for managed_node3/copy 9704 1726776684.13240: done queuing things up, now waiting for results queue to drain 9704 1726776684.13242: waiting for pending results... 10972 1726776684.13421: running TaskExecutor() for managed_node3/TASK: Ensure kernel_settings is not in active_profile 10972 1726776684.13539: in run() - task 120fa90a-8a95-fee7-0cea-00000000058e 10972 1726776684.13555: variable 'ansible_search_path' from source: unknown 10972 1726776684.13562: variable 'ansible_search_path' from source: unknown 10972 1726776684.13591: calling self._execute() 10972 1726776684.13665: variable 'ansible_host' from source: host vars for 'managed_node3' 10972 1726776684.13673: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10972 1726776684.13682: variable 'omit' from source: magic vars 10972 1726776684.13761: variable 'omit' from source: magic vars 10972 1726776684.13792: variable 'omit' from source: magic vars 10972 1726776684.13814: variable '__active_profile' from source: task vars 10972 1726776684.14033: variable '__active_profile' from source: task vars 10972 1726776684.14176: variable '__cur_profile' from source: task vars 10972 1726776684.14278: '/usr/local/lib/python3.12/site-packages/ansible/plugins/filter/__init__' skipped due to reserved name 10972 1726776684.16038: Loading FilterModule 'core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py 10972 1726776684.16085: Loading FilterModule 'encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py 10972 1726776684.16114: Loading FilterModule 'mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py 10972 1726776684.16155: Loading FilterModule 'urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py 10972 1726776684.16180: Loading FilterModule 'urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py 10972 1726776684.16242: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.core' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/core.py (found_in_cache=True, class_only=False) 10972 1726776684.16266: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.encryption' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/encryption.py (found_in_cache=True, class_only=False) 10972 1726776684.16286: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.mathstuff' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/mathstuff.py (found_in_cache=True, class_only=False) 10972 1726776684.16313: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urls' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urls.py (found_in_cache=True, class_only=False) 10972 1726776684.16325: Loading FilterModule 'ansible_collections.ansible.builtin.plugins.filter.urlsplit' from /usr/local/lib/python3.12/site-packages/ansible/plugins/filter/urlsplit.py (found_in_cache=True, class_only=False) 10972 1726776684.16401: variable '__kernel_settings_tuned_current_profile' from source: set_fact 10972 1726776684.16441: variable '__kernel_settings_tuned_active_profile' from source: role '' exported vars 10972 1726776684.16493: variable '__kernel_settings_tuned_active_profile' from source: role '' exported vars 10972 1726776684.16543: variable '__kernel_settings_tuned_dir' from source: role '' exported vars 10972 1726776684.16567: variable 'omit' from source: magic vars 10972 1726776684.16588: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10972 1726776684.16608: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10972 1726776684.16625: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10972 1726776684.16639: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10972 1726776684.16649: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10972 1726776684.16674: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10972 1726776684.16679: variable 'ansible_host' from source: host vars for 'managed_node3' 10972 1726776684.16683: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10972 1726776684.16749: Set connection var ansible_module_compression to ZIP_DEFLATED 10972 1726776684.16760: Set connection var ansible_shell_type to sh 10972 1726776684.16768: Set connection var ansible_timeout to 10 10972 1726776684.16773: Set connection var ansible_pipelining to False 10972 1726776684.16780: Set connection var ansible_shell_executable to /bin/sh 10972 1726776684.16782: Set connection var ansible_connection to ssh 10972 1726776684.16796: variable 'ansible_shell_executable' from source: unknown 10972 1726776684.16799: variable 'ansible_connection' from source: unknown 10972 1726776684.16801: variable 'ansible_module_compression' from source: unknown 10972 1726776684.16803: variable 'ansible_shell_type' from source: unknown 10972 1726776684.16805: variable 'ansible_shell_executable' from source: unknown 10972 1726776684.16807: variable 'ansible_host' from source: host vars for 'managed_node3' 10972 1726776684.16809: variable 'ansible_pipelining' from source: unknown 10972 1726776684.16811: variable 'ansible_timeout' from source: unknown 10972 1726776684.16813: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10972 1726776684.16875: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10972 1726776684.16885: variable 'omit' from source: magic vars 10972 1726776684.16890: starting attempt loop 10972 1726776684.16893: running the handler 10972 1726776684.16901: _low_level_execute_command(): starting 10972 1726776684.16906: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10972 1726776684.19196: stdout chunk (state=2): >>>/root <<< 10972 1726776684.19321: stderr chunk (state=3): >>><<< 10972 1726776684.19332: stdout chunk (state=3): >>><<< 10972 1726776684.19353: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10972 1726776684.19369: _low_level_execute_command(): starting 10972 1726776684.19375: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410 `" && echo ansible-tmp-1726776684.1936386-10972-78526774020410="` echo /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410 `" ) && sleep 0' 10972 1726776684.21850: stdout chunk (state=2): >>>ansible-tmp-1726776684.1936386-10972-78526774020410=/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410 <<< 10972 1726776684.21974: stderr chunk (state=3): >>><<< 10972 1726776684.21981: stdout chunk (state=3): >>><<< 10972 1726776684.21999: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776684.1936386-10972-78526774020410=/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410 , stderr= 10972 1726776684.22069: variable 'ansible_module_compression' from source: unknown 10972 1726776684.22114: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10972 1726776684.22147: variable 'ansible_facts' from source: unknown 10972 1726776684.22213: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_stat.py 10972 1726776684.22300: Sending initial data 10972 1726776684.22307: Sent initial data (151 bytes) 10972 1726776684.24790: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpo_m_nnfb /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_stat.py <<< 10972 1726776684.25741: stderr chunk (state=3): >>><<< 10972 1726776684.25747: stdout chunk (state=3): >>><<< 10972 1726776684.25769: done transferring module to remote 10972 1726776684.25779: _low_level_execute_command(): starting 10972 1726776684.25784: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/ /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_stat.py && sleep 0' 10972 1726776684.28111: stderr chunk (state=2): >>><<< 10972 1726776684.28118: stdout chunk (state=2): >>><<< 10972 1726776684.28134: _low_level_execute_command() done: rc=0, stdout=, stderr= 10972 1726776684.28138: _low_level_execute_command(): starting 10972 1726776684.28143: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_stat.py && sleep 0' 10972 1726776684.44070: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 30, "inode": 142606531, "dev": 51713, "nlink": 1, "atime": 1726776684.08257, "mtime": 1726776681.5245395, "ctime": 1726776681.5245395, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "mimetype": "text/plain", "charset": "us-ascii", "version": "3590528432", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10972 1726776684.45195: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10972 1726776684.45242: stderr chunk (state=3): >>><<< 10972 1726776684.45248: stdout chunk (state=3): >>><<< 10972 1726776684.45266: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/active_profile", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 30, "inode": 142606531, "dev": 51713, "nlink": 1, "atime": 1726776684.08257, "mtime": 1726776681.5245395, "ctime": 1726776681.5245395, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "a79569d3860cb6a066e0e92c8b22ffd0e8796bfd", "mimetype": "text/plain", "charset": "us-ascii", "version": "3590528432", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/active_profile", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10972 1726776684.45306: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/active_profile', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10972 1726776684.45389: Sending initial data 10972 1726776684.45396: Sent initial data (140 bytes) 10972 1726776684.47904: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp_pvc26io /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source <<< 10972 1726776684.48153: stderr chunk (state=3): >>><<< 10972 1726776684.48161: stdout chunk (state=3): >>><<< 10972 1726776684.48180: _low_level_execute_command(): starting 10972 1726776684.48186: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/ /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source && sleep 0' 10972 1726776684.50497: stderr chunk (state=2): >>><<< 10972 1726776684.50505: stdout chunk (state=2): >>><<< 10972 1726776684.50518: _low_level_execute_command() done: rc=0, stdout=, stderr= 10972 1726776684.50539: variable 'ansible_module_compression' from source: unknown 10972 1726776684.50575: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.copy-ZIP_DEFLATED 10972 1726776684.50593: variable 'ansible_facts' from source: unknown 10972 1726776684.50652: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_copy.py 10972 1726776684.50739: Sending initial data 10972 1726776684.50746: Sent initial data (151 bytes) 10972 1726776684.53202: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpi4znicsw /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_copy.py <<< 10972 1726776684.54367: stderr chunk (state=3): >>><<< 10972 1726776684.54376: stdout chunk (state=3): >>><<< 10972 1726776684.54400: done transferring module to remote 10972 1726776684.54411: _low_level_execute_command(): starting 10972 1726776684.54416: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/ /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_copy.py && sleep 0' 10972 1726776684.56875: stderr chunk (state=2): >>><<< 10972 1726776684.56884: stdout chunk (state=2): >>><<< 10972 1726776684.56899: _low_level_execute_command() done: rc=0, stdout=, stderr= 10972 1726776684.56904: _low_level_execute_command(): starting 10972 1726776684.56909: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/AnsiballZ_copy.py && sleep 0' 10972 1726776684.73306: stdout chunk (state=2): >>> {"dest": "/etc/tuned/active_profile", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source", "md5sum": "9a561d913bcdb5a659ec2dd035975a8e", "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 14, "invocation": {"module_args": {"dest": "/etc/tuned/active_profile", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source", "_original_basename": "tmp_pvc26io", "follow": false, "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10972 1726776684.74491: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10972 1726776684.74538: stderr chunk (state=3): >>><<< 10972 1726776684.74545: stdout chunk (state=3): >>><<< 10972 1726776684.74563: _low_level_execute_command() done: rc=0, stdout= {"dest": "/etc/tuned/active_profile", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source", "md5sum": "9a561d913bcdb5a659ec2dd035975a8e", "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 14, "invocation": {"module_args": {"dest": "/etc/tuned/active_profile", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source", "_original_basename": "tmp_pvc26io", "follow": false, "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10972 1726776684.74588: done with _execute_module (ansible.legacy.copy, {'dest': '/etc/tuned/active_profile', 'mode': '0600', 'src': '/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source', '_original_basename': 'tmp_pvc26io', 'follow': False, 'checksum': '633f07e1b5698d04352d5dca735869bf2fe77897', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10972 1726776684.74599: _low_level_execute_command(): starting 10972 1726776684.74604: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/ > /dev/null 2>&1 && sleep 0' 10972 1726776684.76969: stderr chunk (state=2): >>><<< 10972 1726776684.76977: stdout chunk (state=2): >>><<< 10972 1726776684.76990: _low_level_execute_command() done: rc=0, stdout=, stderr= 10972 1726776684.76998: handler run complete 10972 1726776684.77018: attempt loop complete, returning result 10972 1726776684.77022: _execute() done 10972 1726776684.77025: dumping result to json 10972 1726776684.77033: done dumping result, returning 10972 1726776684.77039: done running TaskExecutor() for managed_node3/TASK: Ensure kernel_settings is not in active_profile [120fa90a-8a95-fee7-0cea-00000000058e] 10972 1726776684.77045: sending task result for task 120fa90a-8a95-fee7-0cea-00000000058e 10972 1726776684.77076: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000058e 10972 1726776684.77080: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "633f07e1b5698d04352d5dca735869bf2fe77897", "dest": "/etc/tuned/active_profile", "gid": 0, "group": "root", "md5sum": "9a561d913bcdb5a659ec2dd035975a8e", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:tuned_rw_etc_t:s0", "size": 14, "src": "/root/.ansible/tmp/ansible-tmp-1726776684.1936386-10972-78526774020410/source", "state": "file", "uid": 0 } 9704 1726776684.77218: no more pending results, returning what we have 9704 1726776684.77221: results queue empty 9704 1726776684.77222: checking for any_errors_fatal 9704 1726776684.77231: done checking for any_errors_fatal 9704 1726776684.77232: checking for max_fail_percentage 9704 1726776684.77233: done checking for max_fail_percentage 9704 1726776684.77234: checking to see if all hosts have failed and the running result is not ok 9704 1726776684.77234: done checking to see if all hosts have failed 9704 1726776684.77235: getting the remaining hosts for this loop 9704 1726776684.77236: done getting the remaining hosts for this loop 9704 1726776684.77239: getting the next task for host managed_node3 9704 1726776684.77245: done getting next task for host managed_node3 9704 1726776684.77247: ^ task is: TASK: Set profile_mode to auto 9704 1726776684.77249: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=4, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776684.77252: getting variables 9704 1726776684.77254: in VariableManager get_vars() 9704 1726776684.77289: Calling all_inventory to load vars for managed_node3 9704 1726776684.77292: Calling groups_inventory to load vars for managed_node3 9704 1726776684.77294: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776684.77303: Calling all_plugins_play to load vars for managed_node3 9704 1726776684.77305: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776684.77308: Calling groups_plugins_play to load vars for managed_node3 9704 1726776684.77437: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776684.77555: done with get_vars() 9704 1726776684.77565: done getting variables 9704 1726776684.77608: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Set profile_mode to auto] ************************************************ task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:57 Thursday 19 September 2024 16:11:24 -0400 (0:00:00.646) 0:00:37.705 **** 9704 1726776684.77632: entering _queue_task() for managed_node3/copy 9704 1726776684.77803: worker is 1 (out of 1 available) 9704 1726776684.77818: exiting _queue_task() for managed_node3/copy 9704 1726776684.77831: done queuing things up, now waiting for results queue to drain 9704 1726776684.77833: waiting for pending results... 10999 1726776684.77954: running TaskExecutor() for managed_node3/TASK: Set profile_mode to auto 10999 1726776684.78062: in run() - task 120fa90a-8a95-fee7-0cea-00000000058f 10999 1726776684.78077: variable 'ansible_search_path' from source: unknown 10999 1726776684.78081: variable 'ansible_search_path' from source: unknown 10999 1726776684.78108: calling self._execute() 10999 1726776684.78178: variable 'ansible_host' from source: host vars for 'managed_node3' 10999 1726776684.78186: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10999 1726776684.78195: variable 'omit' from source: magic vars 10999 1726776684.78274: variable 'omit' from source: magic vars 10999 1726776684.78304: variable 'omit' from source: magic vars 10999 1726776684.78325: variable '__kernel_settings_tuned_profile_mode' from source: role '' exported vars 10999 1726776684.78535: variable '__kernel_settings_tuned_profile_mode' from source: role '' exported vars 10999 1726776684.78593: variable '__kernel_settings_tuned_dir' from source: role '' exported vars 10999 1726776684.78681: variable 'omit' from source: magic vars 10999 1726776684.78714: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 10999 1726776684.78742: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 10999 1726776684.78761: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 10999 1726776684.78775: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10999 1726776684.78784: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 10999 1726776684.78808: variable 'inventory_hostname' from source: host vars for 'managed_node3' 10999 1726776684.78813: variable 'ansible_host' from source: host vars for 'managed_node3' 10999 1726776684.78817: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10999 1726776684.78886: Set connection var ansible_module_compression to ZIP_DEFLATED 10999 1726776684.78894: Set connection var ansible_shell_type to sh 10999 1726776684.78902: Set connection var ansible_timeout to 10 10999 1726776684.78908: Set connection var ansible_pipelining to False 10999 1726776684.78914: Set connection var ansible_shell_executable to /bin/sh 10999 1726776684.78916: Set connection var ansible_connection to ssh 10999 1726776684.78932: variable 'ansible_shell_executable' from source: unknown 10999 1726776684.78936: variable 'ansible_connection' from source: unknown 10999 1726776684.78938: variable 'ansible_module_compression' from source: unknown 10999 1726776684.78940: variable 'ansible_shell_type' from source: unknown 10999 1726776684.78941: variable 'ansible_shell_executable' from source: unknown 10999 1726776684.78943: variable 'ansible_host' from source: host vars for 'managed_node3' 10999 1726776684.78945: variable 'ansible_pipelining' from source: unknown 10999 1726776684.78947: variable 'ansible_timeout' from source: unknown 10999 1726776684.78949: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 10999 1726776684.79033: Loading ActionModule 'copy' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/copy.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 10999 1726776684.79042: variable 'omit' from source: magic vars 10999 1726776684.79047: starting attempt loop 10999 1726776684.79049: running the handler 10999 1726776684.79058: _low_level_execute_command(): starting 10999 1726776684.79064: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 10999 1726776684.81326: stdout chunk (state=2): >>>/root <<< 10999 1726776684.81444: stderr chunk (state=3): >>><<< 10999 1726776684.81450: stdout chunk (state=3): >>><<< 10999 1726776684.81467: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 10999 1726776684.81480: _low_level_execute_command(): starting 10999 1726776684.81486: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182 `" && echo ansible-tmp-1726776684.8147478-10999-64732855484182="` echo /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182 `" ) && sleep 0' 10999 1726776684.83931: stdout chunk (state=2): >>>ansible-tmp-1726776684.8147478-10999-64732855484182=/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182 <<< 10999 1726776684.84050: stderr chunk (state=3): >>><<< 10999 1726776684.84057: stdout chunk (state=3): >>><<< 10999 1726776684.84073: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776684.8147478-10999-64732855484182=/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182 , stderr= 10999 1726776684.84140: variable 'ansible_module_compression' from source: unknown 10999 1726776684.84184: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.stat-ZIP_DEFLATED 10999 1726776684.84210: variable 'ansible_facts' from source: unknown 10999 1726776684.84282: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_stat.py 10999 1726776684.84362: Sending initial data 10999 1726776684.84369: Sent initial data (151 bytes) 10999 1726776684.86861: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpmc0tkdck /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_stat.py <<< 10999 1726776684.87840: stderr chunk (state=3): >>><<< 10999 1726776684.87847: stdout chunk (state=3): >>><<< 10999 1726776684.87864: done transferring module to remote 10999 1726776684.87876: _low_level_execute_command(): starting 10999 1726776684.87881: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/ /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_stat.py && sleep 0' 10999 1726776684.90399: stderr chunk (state=2): >>><<< 10999 1726776684.90405: stdout chunk (state=2): >>><<< 10999 1726776684.90417: _low_level_execute_command() done: rc=0, stdout=, stderr= 10999 1726776684.90421: _low_level_execute_command(): starting 10999 1726776684.90426: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_stat.py && sleep 0' 10999 1726776685.06300: stdout chunk (state=2): >>> {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 7, "inode": 148897989, "dev": 51713, "nlink": 1, "atime": 1726776664.3133166, "mtime": 1726776681.5245395, "ctime": 1726776681.5245395, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "mimetype": "text/plain", "charset": "us-ascii", "version": "1846195820", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} <<< 10999 1726776685.07408: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10999 1726776685.07464: stderr chunk (state=3): >>><<< 10999 1726776685.07472: stdout chunk (state=3): >>><<< 10999 1726776685.07491: _low_level_execute_command() done: rc=0, stdout= {"changed": false, "stat": {"exists": true, "path": "/etc/tuned/profile_mode", "mode": "0600", "isdir": false, "ischr": false, "isblk": false, "isreg": true, "isfifo": false, "islnk": false, "issock": false, "uid": 0, "gid": 0, "size": 7, "inode": 148897989, "dev": 51713, "nlink": 1, "atime": 1726776664.3133166, "mtime": 1726776681.5245395, "ctime": 1726776681.5245395, "wusr": true, "rusr": true, "xusr": false, "wgrp": false, "rgrp": false, "xgrp": false, "woth": false, "roth": false, "xoth": false, "isuid": false, "isgid": false, "blocks": 8, "block_size": 4096, "device_type": 0, "readable": true, "writeable": true, "executable": false, "pw_name": "root", "gr_name": "root", "checksum": "3ef9f23deed2e23d3ef2b88b842fb882313e15ce", "mimetype": "text/plain", "charset": "us-ascii", "version": "1846195820", "attributes": [], "attr_flags": ""}, "invocation": {"module_args": {"path": "/etc/tuned/profile_mode", "follow": false, "get_checksum": true, "checksum_algorithm": "sha1", "get_mime": true, "get_attributes": true}}} , stderr=Shared connection to 10.31.8.186 closed. 10999 1726776685.07544: done with _execute_module (ansible.legacy.stat, {'path': '/etc/tuned/profile_mode', 'follow': False, 'get_checksum': True, 'checksum_algorithm': 'sha1', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.stat', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10999 1726776685.07795: Sending initial data 10999 1726776685.07802: Sent initial data (140 bytes) 10999 1726776685.10331: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpuzw97z9v /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source <<< 10999 1726776685.10912: stderr chunk (state=3): >>><<< 10999 1726776685.10919: stdout chunk (state=3): >>><<< 10999 1726776685.10939: _low_level_execute_command(): starting 10999 1726776685.10946: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/ /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source && sleep 0' 10999 1726776685.13240: stderr chunk (state=2): >>><<< 10999 1726776685.13247: stdout chunk (state=2): >>><<< 10999 1726776685.13261: _low_level_execute_command() done: rc=0, stdout=, stderr= 10999 1726776685.13280: variable 'ansible_module_compression' from source: unknown 10999 1726776685.13312: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.copy-ZIP_DEFLATED 10999 1726776685.13331: variable 'ansible_facts' from source: unknown 10999 1726776685.13390: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_copy.py 10999 1726776685.13471: Sending initial data 10999 1726776685.13477: Sent initial data (151 bytes) 10999 1726776685.15849: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmpv23b62b4 /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_copy.py <<< 10999 1726776685.16831: stderr chunk (state=3): >>><<< 10999 1726776685.16838: stdout chunk (state=3): >>><<< 10999 1726776685.16856: done transferring module to remote 10999 1726776685.16865: _low_level_execute_command(): starting 10999 1726776685.16870: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/ /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_copy.py && sleep 0' 10999 1726776685.19134: stderr chunk (state=2): >>><<< 10999 1726776685.19142: stdout chunk (state=2): >>><<< 10999 1726776685.19155: _low_level_execute_command() done: rc=0, stdout=, stderr= 10999 1726776685.19159: _low_level_execute_command(): starting 10999 1726776685.19165: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/AnsiballZ_copy.py && sleep 0' 10999 1726776685.35913: stdout chunk (state=2): >>> {"dest": "/etc/tuned/profile_mode", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source", "md5sum": "451e20aff0f489cd2f7d4d73533aa961", "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 5, "invocation": {"module_args": {"dest": "/etc/tuned/profile_mode", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source", "_original_basename": "tmpuzw97z9v", "follow": false, "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} <<< 10999 1726776685.37299: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 10999 1726776685.37313: stdout chunk (state=3): >>><<< 10999 1726776685.37325: stderr chunk (state=3): >>><<< 10999 1726776685.37344: _low_level_execute_command() done: rc=0, stdout= {"dest": "/etc/tuned/profile_mode", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source", "md5sum": "451e20aff0f489cd2f7d4d73533aa961", "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "changed": true, "uid": 0, "gid": 0, "owner": "root", "group": "root", "mode": "0600", "state": "file", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 5, "invocation": {"module_args": {"dest": "/etc/tuned/profile_mode", "mode": "0600", "src": "/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source", "_original_basename": "tmpuzw97z9v", "follow": false, "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "backup": false, "force": true, "unsafe_writes": false, "content": null, "validate": null, "directory_mode": null, "remote_src": null, "local_follow": null, "owner": null, "group": null, "seuser": null, "serole": null, "selevel": null, "setype": null, "attributes": null}}} , stderr=Shared connection to 10.31.8.186 closed. 10999 1726776685.37382: done with _execute_module (ansible.legacy.copy, {'dest': '/etc/tuned/profile_mode', 'mode': '0600', 'src': '/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source', '_original_basename': 'tmpuzw97z9v', 'follow': False, 'checksum': '43683f4e92c48be4b00ddd86e011a4f27fcdbeb5', '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.copy', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 10999 1726776685.37394: _low_level_execute_command(): starting 10999 1726776685.37400: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/ > /dev/null 2>&1 && sleep 0' 10999 1726776685.40114: stderr chunk (state=2): >>><<< 10999 1726776685.40125: stdout chunk (state=2): >>><<< 10999 1726776685.40143: _low_level_execute_command() done: rc=0, stdout=, stderr= 10999 1726776685.40152: handler run complete 10999 1726776685.40172: attempt loop complete, returning result 10999 1726776685.40176: _execute() done 10999 1726776685.40180: dumping result to json 10999 1726776685.40187: done dumping result, returning 10999 1726776685.40195: done running TaskExecutor() for managed_node3/TASK: Set profile_mode to auto [120fa90a-8a95-fee7-0cea-00000000058f] 10999 1726776685.40201: sending task result for task 120fa90a-8a95-fee7-0cea-00000000058f 10999 1726776685.40234: done sending task result for task 120fa90a-8a95-fee7-0cea-00000000058f 10999 1726776685.40238: WORKER PROCESS EXITING changed: [managed_node3] => { "changed": true, "checksum": "43683f4e92c48be4b00ddd86e011a4f27fcdbeb5", "dest": "/etc/tuned/profile_mode", "gid": 0, "group": "root", "md5sum": "451e20aff0f489cd2f7d4d73533aa961", "mode": "0600", "owner": "root", "secontext": "system_u:object_r:tuned_etc_t:s0", "size": 5, "src": "/root/.ansible/tmp/ansible-tmp-1726776684.8147478-10999-64732855484182/source", "state": "file", "uid": 0 } 9704 1726776685.40375: no more pending results, returning what we have 9704 1726776685.40378: results queue empty 9704 1726776685.40379: checking for any_errors_fatal 9704 1726776685.40389: done checking for any_errors_fatal 9704 1726776685.40392: checking for max_fail_percentage 9704 1726776685.40394: done checking for max_fail_percentage 9704 1726776685.40394: checking to see if all hosts have failed and the running result is not ok 9704 1726776685.40395: done checking to see if all hosts have failed 9704 1726776685.40395: getting the remaining hosts for this loop 9704 1726776685.40397: done getting the remaining hosts for this loop 9704 1726776685.40400: getting the next task for host managed_node3 9704 1726776685.40405: done getting next task for host managed_node3 9704 1726776685.40408: ^ task is: TASK: Restart tuned 9704 1726776685.40410: ^ state is: HOST STATE: block=2, task=20, rescue=0, always=3, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (HOST STATE: block=0, task=6, rescue=0, always=5, handlers=0, run_state=3, fail_state=0, pre_flushing_run_state=None, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False), did rescue? False, did start at task? False 9704 1726776685.40414: getting variables 9704 1726776685.40415: in VariableManager get_vars() 9704 1726776685.40449: Calling all_inventory to load vars for managed_node3 9704 1726776685.40452: Calling groups_inventory to load vars for managed_node3 9704 1726776685.40454: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776685.40463: Calling all_plugins_play to load vars for managed_node3 9704 1726776685.40466: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776685.40468: Calling groups_plugins_play to load vars for managed_node3 9704 1726776685.40632: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776685.40754: done with get_vars() 9704 1726776685.40765: done getting variables 9704 1726776685.40822: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=True) TASK [Restart tuned] *********************************************************** task path: /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:64 Thursday 19 September 2024 16:11:25 -0400 (0:00:00.632) 0:00:38.337 **** 9704 1726776685.40851: entering _queue_task() for managed_node3/service 9704 1726776685.41070: worker is 1 (out of 1 available) 9704 1726776685.41084: exiting _queue_task() for managed_node3/service 9704 1726776685.41095: done queuing things up, now waiting for results queue to drain 9704 1726776685.41096: waiting for pending results... 11032 1726776685.41315: running TaskExecutor() for managed_node3/TASK: Restart tuned 11032 1726776685.41454: in run() - task 120fa90a-8a95-fee7-0cea-000000000590 11032 1726776685.41473: variable 'ansible_search_path' from source: unknown 11032 1726776685.41478: variable 'ansible_search_path' from source: unknown 11032 1726776685.41515: variable '__kernel_settings_services' from source: include_vars 11032 1726776685.41773: variable '__kernel_settings_services' from source: include_vars 11032 1726776685.41831: variable 'omit' from source: magic vars 11032 1726776685.41922: variable 'ansible_host' from source: host vars for 'managed_node3' 11032 1726776685.41934: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 11032 1726776685.41943: variable 'omit' from source: magic vars 11032 1726776685.41994: variable 'omit' from source: magic vars 11032 1726776685.42019: variable 'omit' from source: magic vars 11032 1726776685.42052: variable 'item' from source: unknown 11032 1726776685.42108: variable 'item' from source: unknown 11032 1726776685.42133: variable 'omit' from source: magic vars 11032 1726776685.42163: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/connection 11032 1726776685.42189: Loading Connection 'ssh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/connection/ssh.py (found_in_cache=True, class_only=False) 11032 1726776685.42207: trying /usr/local/lib/python3.12/site-packages/ansible/plugins/shell 11032 1726776685.42222: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 11032 1726776685.42236: Loading ShellModule 'sh' from /usr/local/lib/python3.12/site-packages/ansible/plugins/shell/sh.py (found_in_cache=True, class_only=False) 11032 1726776685.42259: variable 'inventory_hostname' from source: host vars for 'managed_node3' 11032 1726776685.42264: variable 'ansible_host' from source: host vars for 'managed_node3' 11032 1726776685.42268: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 11032 1726776685.42335: Set connection var ansible_module_compression to ZIP_DEFLATED 11032 1726776685.42343: Set connection var ansible_shell_type to sh 11032 1726776685.42351: Set connection var ansible_timeout to 10 11032 1726776685.42356: Set connection var ansible_pipelining to False 11032 1726776685.42364: Set connection var ansible_shell_executable to /bin/sh 11032 1726776685.42367: Set connection var ansible_connection to ssh 11032 1726776685.42381: variable 'ansible_shell_executable' from source: unknown 11032 1726776685.42384: variable 'ansible_connection' from source: unknown 11032 1726776685.42388: variable 'ansible_module_compression' from source: unknown 11032 1726776685.42392: variable 'ansible_shell_type' from source: unknown 11032 1726776685.42395: variable 'ansible_shell_executable' from source: unknown 11032 1726776685.42399: variable 'ansible_host' from source: host vars for 'managed_node3' 11032 1726776685.42403: variable 'ansible_pipelining' from source: unknown 11032 1726776685.42406: variable 'ansible_timeout' from source: unknown 11032 1726776685.42410: variable 'ansible_ssh_extra_args' from source: host vars for 'managed_node3' 11032 1726776685.42501: Loading ActionModule 'service' from /usr/local/lib/python3.12/site-packages/ansible/plugins/action/service.py (searched paths: /usr/local/lib/python3.12/site-packages/ansible/plugins/action/__pycache__:/usr/local/lib/python3.12/site-packages/ansible/plugins/action) (found_in_cache=True, class_only=False) 11032 1726776685.42513: variable 'omit' from source: magic vars 11032 1726776685.42519: starting attempt loop 11032 1726776685.42523: running the handler 11032 1726776685.42585: variable 'ansible_facts' from source: unknown 11032 1726776685.42674: _low_level_execute_command(): starting 11032 1726776685.42683: _low_level_execute_command(): executing: /bin/sh -c 'echo ~ && sleep 0' 11032 1726776685.44996: stdout chunk (state=2): >>>/root <<< 11032 1726776685.45113: stderr chunk (state=3): >>><<< 11032 1726776685.45122: stdout chunk (state=3): >>><<< 11032 1726776685.45140: _low_level_execute_command() done: rc=0, stdout=/root , stderr= 11032 1726776685.45153: _low_level_execute_command(): starting 11032 1726776685.45159: _low_level_execute_command(): executing: /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819 `" && echo ansible-tmp-1726776685.451478-11032-168933111548819="` echo /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819 `" ) && sleep 0' 11032 1726776685.47622: stdout chunk (state=2): >>>ansible-tmp-1726776685.451478-11032-168933111548819=/root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819 <<< 11032 1726776685.47745: stderr chunk (state=3): >>><<< 11032 1726776685.47751: stdout chunk (state=3): >>><<< 11032 1726776685.47765: _low_level_execute_command() done: rc=0, stdout=ansible-tmp-1726776685.451478-11032-168933111548819=/root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819 , stderr= 11032 1726776685.47787: variable 'ansible_module_compression' from source: unknown 11032 1726776685.47831: ANSIBALLZ: using cached module: /root/.ansible/tmp/ansible-local-9704nc432w9e/ansiballz_cache/ansible.modules.systemd-ZIP_DEFLATED 11032 1726776685.47874: variable 'ansible_facts' from source: unknown 11032 1726776685.48030: transferring module to remote /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/AnsiballZ_systemd.py 11032 1726776685.48134: Sending initial data 11032 1726776685.48141: Sent initial data (154 bytes) 11032 1726776685.50633: stdout chunk (state=3): >>>sftp> put /root/.ansible/tmp/ansible-local-9704nc432w9e/tmp7u7nul2b /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/AnsiballZ_systemd.py <<< 11032 1726776685.52502: stderr chunk (state=3): >>><<< 11032 1726776685.52510: stdout chunk (state=3): >>><<< 11032 1726776685.52531: done transferring module to remote 11032 1726776685.52544: _low_level_execute_command(): starting 11032 1726776685.52549: _low_level_execute_command(): executing: /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/ /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/AnsiballZ_systemd.py && sleep 0' 11032 1726776685.54935: stderr chunk (state=2): >>><<< 11032 1726776685.54944: stdout chunk (state=2): >>><<< 11032 1726776685.54956: _low_level_execute_command() done: rc=0, stdout=, stderr= 11032 1726776685.54961: _low_level_execute_command(): starting 11032 1726776685.54967: _low_level_execute_command(): executing: /bin/sh -c '/usr/libexec/platform-python /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/AnsiballZ_systemd.py && sleep 0' 11032 1726776685.82959: stdout chunk (state=2): >>> {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "15002", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15376384", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "Memo<<< 11032 1726776685.83000: stdout chunk (state=3): >>>ryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} <<< 11032 1726776685.84684: stderr chunk (state=3): >>>Shared connection to 10.31.8.186 closed. <<< 11032 1726776685.84732: stderr chunk (state=3): >>><<< 11032 1726776685.84740: stdout chunk (state=3): >>><<< 11032 1726776685.84761: _low_level_execute_command() done: rc=0, stdout= {"name": "tuned", "changed": false, "status": {"Type": "dbus", "Restart": "no", "PIDFile": "/run/tuned/tuned.pid", "NotifyAccess": "none", "RestartUSec": "100ms", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "RuntimeMaxUSec": "infinity", "WatchdogUSec": "0", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "PermissionsStartOnly": "no", "RootDirectoryStartOnly": "no", "RemainAfterExit": "no", "GuessMainPID": "yes", "MainPID": "15002", "ControlPID": "0", "BusName": "com.redhat.tuned", "FileDescriptorStoreMax": "0", "NFileDescriptorStore": "0", "StatusErrno": "0", "Result": "success", "UID": "[not set]", "GID": "[not set]", "NRestarts": "0", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainCode": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "Slice": "system.slice", "ControlGroup": "/system.slice/tuned.service", "MemoryCurrent": "15376384", "CPUUsageNSec": "[not set]", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "TasksCurrent": "4", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "Delegate": "no", "CPUAccounting": "no", "CPUWeight": "[not set]", "StartupCPUWeight": "[not set]", "CPUShares": "[not set]", "StartupCPUShares": "[not set]", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "AllowedCPUs": "", "AllowedMemoryNodes": "", "IOAccounting": "no", "IOWeight": "[not set]", "StartupIOWeight": "[not set]", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "StartupBlockIOWeight": "[not set]", "MemoryAccounting": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "MemoryMin": "0", "MemoryLow": "0", "MemoryHigh": "infinity", "MemoryMax": "infinity", "MemorySwapMax": "infinity", "MemoryLimit": "infinity", "DevicePolicy": "auto", "TasksAccounting": "yes", "TasksMax": "22405", "IPAccounting": "no", "UMask": "0022", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "OOMScoreAdjust": "0", "Nice": "0", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "NUMAPolicy": "n/a", "NUMAMask": "", "TimerSlackNSec": "50000", "CPUSchedulingResetOnFork": "no", "NonBlocking": "no", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StandardError": "inherit", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "SyslogPriority": "30", "SyslogLevelPrefix": "yes", "SyslogLevel": "6", "SyslogFacility": "3", "LogLevelMax": "-1", "LogRateLimitIntervalUSec": "0", "LogRateLimitBurst": "0", "SecureBits": "0", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "AmbientCapabilities": "", "DynamicUser": "no", "RemoveIPC": "no", "MountFlags": "", "PrivateTmp": "no", "PrivateDevices": "no", "ProtectKernelTunables": "no", "ProtectKernelModules": "no", "ProtectControlGroups": "no", "PrivateNetwork": "no", "PrivateUsers": "no", "PrivateMounts": "no", "ProtectHome": "no", "ProtectSystem": "no", "SameProcessGroup": "no", "UtmpMode": "init", "IgnoreSIGPIPE": "yes", "NoNewPrivileges": "no", "SystemCallErrorNumber": "0", "LockPersonality": "no", "RuntimeDirectoryPreserve": "no", "RuntimeDirectoryMode": "0755", "StateDirectoryMode": "0755", "CacheDirectoryMode": "0755", "LogsDirectoryMode": "0755", "ConfigurationDirectoryMode": "0755", "MemoryDenyWriteExecute": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "RestrictNamespaces": "no", "MountAPIVFS": "no", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "SendSIGKILL": "yes", "SendSIGHUP": "no", "Id": "tuned.service", "Names": "tuned.service", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "WantedBy": "multi-user.target", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "Before": "multi-user.target shutdown.target", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "Description": "Dynamic System Tuning Daemon", "LoadState": "loaded", "ActiveState": "active", "FreezerState": "running", "SubState": "running", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "UnitFileState": "enabled", "UnitFilePreset": "enabled", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "CanStart": "yes", "CanStop": "yes", "CanReload": "no", "CanIsolate": "no", "CanFreeze": "yes", "StopWhenUnneeded": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "AllowIsolate": "no", "DefaultDependencies": "yes", "OnFailureJobMode": "replace", "IgnoreOnIsolate": "no", "NeedDaemonReload": "no", "JobTimeoutUSec": "infinity", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "ConditionResult": "yes", "AssertResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Transient": "no", "Perpetual": "no", "StartLimitIntervalUSec": "10s", "StartLimitBurst": "5", "StartLimitAction": "none", "FailureAction": "none", "SuccessAction": "none", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "CollectMode": "inactive"}, "enabled": true, "state": "started", "invocation": {"module_args": {"name": "tuned", "state": "started", "enabled": true, "daemon_reload": false, "daemon_reexec": false, "scope": "system", "no_block": false, "force": null, "masked": null}}} , stderr=Shared connection to 10.31.8.186 closed. 11032 1726776685.85167: done with _execute_module (ansible.legacy.systemd, {'name': 'tuned', 'state': 'started', 'enabled': True, '_ansible_check_mode': False, '_ansible_no_log': False, '_ansible_debug': True, '_ansible_diff': False, '_ansible_verbosity': 2, '_ansible_version': '2.16.11', '_ansible_module_name': 'ansible.legacy.systemd', '_ansible_syslog_facility': 'LOG_USER', '_ansible_selinux_special_fs': ['fuse', 'nfs', 'vboxsf', 'ramfs', '9p', 'vfat'], '_ansible_string_conversion_action': 'warn', '_ansible_socket': None, '_ansible_shell_executable': '/bin/sh', '_ansible_keep_remote_files': False, '_ansible_tmpdir': '/root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/', '_ansible_remote_tmp': '~/.ansible/tmp'}) 11032 1726776685.85187: _low_level_execute_command(): starting 11032 1726776685.85193: _low_level_execute_command(): executing: /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1726776685.451478-11032-168933111548819/ > /dev/null 2>&1 && sleep 0' 11032 1726776685.87599: stderr chunk (state=2): >>><<< 11032 1726776685.87611: stdout chunk (state=2): >>><<< 11032 1726776685.87626: _low_level_execute_command() done: rc=0, stdout=, stderr= 11032 1726776685.87636: handler run complete 11032 1726776685.87670: attempt loop complete, returning result 11032 1726776685.87688: variable 'item' from source: unknown 11032 1726776685.87754: variable 'item' from source: unknown ok: [managed_node3] => (item=tuned) => { "ansible_loop_var": "item", "changed": false, "enabled": true, "item": "tuned", "name": "tuned", "state": "started", "status": { "ActiveEnterTimestamp": "Thu 2024-09-19 16:10:58 EDT", "ActiveEnterTimestampMonotonic": "266070308", "ActiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ActiveExitTimestampMonotonic": "265842818", "ActiveState": "active", "After": "system.slice systemd-journald.socket dbus.service network.target sysinit.target basic.target dbus.socket systemd-sysctl.service polkit.service", "AllowIsolate": "no", "AllowedCPUs": "", "AllowedMemoryNodes": "", "AmbientCapabilities": "", "AssertResult": "yes", "AssertTimestamp": "Thu 2024-09-19 16:10:57 EDT", "AssertTimestampMonotonic": "265928701", "Before": "multi-user.target shutdown.target", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "BusName": "com.redhat.tuned", "CPUAccounting": "no", "CPUAffinity": "", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf", "CollectMode": "inactive", "ConditionResult": "yes", "ConditionTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ConditionTimestampMonotonic": "265928699", "ConfigurationDirectoryMode": "0755", "Conflicts": "auto-cpufreq.service tlp.service power-profiles-daemon.service shutdown.target cpupower.service", "ControlGroup": "/system.slice/tuned.service", "ControlPID": "0", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Dynamic System Tuning Daemon", "DevicePolicy": "auto", "Documentation": "man:tuned(8) man:tuned.conf(5) man:tuned-adm(8)", "DynamicUser": "no", "EffectiveCPUs": "", "EffectiveMemoryNodes": "", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "15002", "ExecMainStartTimestamp": "Thu 2024-09-19 16:10:57 EDT", "ExecMainStartTimestampMonotonic": "265929800", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/sbin/tuned ; argv[]=/usr/sbin/tuned -l -P ; ignore_errors=no ; start_time=[Thu 2024-09-19 16:10:57 EDT] ; stop_time=[n/a] ; pid=15002 ; code=(null) ; status=0/0 }", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FragmentPath": "/usr/lib/systemd/system/tuned.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOSchedulingClass": "0", "IOSchedulingPriority": "0", "IOWeight": "[not set]", "IPAccounting": "no", "IPEgressBytes": "18446744073709551615", "IPEgressPackets": "18446744073709551615", "IPIngressBytes": "18446744073709551615", "IPIngressPackets": "18446744073709551615", "Id": "tuned.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveEnterTimestampMonotonic": "265927725", "InactiveExitTimestamp": "Thu 2024-09-19 16:10:57 EDT", "InactiveExitTimestampMonotonic": "265929843", "InvocationID": "2aa2930e6a8341179bb30611db9523fd", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "0", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "65536", "LimitMEMLOCKSoft": "65536", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "262144", "LimitNOFILESoft": "1024", "LimitNPROC": "14003", "LimitNPROCSoft": "14003", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14003", "LimitSIGPENDINGSoft": "14003", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "15002", "MemoryAccounting": "yes", "MemoryCurrent": "15376384", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MountAPIVFS": "no", "MountFlags": "", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAMask": "", "NUMAPolicy": "n/a", "Names": "tuned.service", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "PIDFile": "/run/tuned/tuned.pid", "PermissionsStartOnly": "no", "Perpetual": "no", "PrivateDevices": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "RemainAfterExit": "no", "RemoveIPC": "no", "Requires": "sysinit.target system.slice dbus.socket dbus.service", "Restart": "no", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system.slice", "StandardError": "inherit", "StandardInput": "null", "StandardInputData": "", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2024-09-19 16:10:58 EDT", "StateChangeTimestampMonotonic": "266070308", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "running", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "0", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "4", "TasksMax": "22405", "TimeoutStartUSec": "1min 30s", "TimeoutStopUSec": "1min 30s", "TimerSlackNSec": "50000", "Transient": "no", "Type": "dbus", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "enabled", "UnitFileState": "enabled", "UtmpMode": "init", "WantedBy": "multi-user.target", "WatchdogTimestamp": "Thu 2024-09-19 16:10:58 EDT", "WatchdogTimestampMonotonic": "266070305", "WatchdogUSec": "0" } } 11032 1726776685.87859: dumping result to json 11032 1726776685.87878: done dumping result, returning 11032 1726776685.87886: done running TaskExecutor() for managed_node3/TASK: Restart tuned [120fa90a-8a95-fee7-0cea-000000000590] 11032 1726776685.87892: sending task result for task 120fa90a-8a95-fee7-0cea-000000000590 11032 1726776685.88004: done sending task result for task 120fa90a-8a95-fee7-0cea-000000000590 11032 1726776685.88008: WORKER PROCESS EXITING 9704 1726776685.88549: no more pending results, returning what we have 9704 1726776685.88552: results queue empty 9704 1726776685.88552: checking for any_errors_fatal 9704 1726776685.88555: done checking for any_errors_fatal 9704 1726776685.88556: checking for max_fail_percentage 9704 1726776685.88559: done checking for max_fail_percentage 9704 1726776685.88559: checking to see if all hosts have failed and the running result is not ok 9704 1726776685.88559: done checking to see if all hosts have failed 9704 1726776685.88560: getting the remaining hosts for this loop 9704 1726776685.88561: done getting the remaining hosts for this loop 9704 1726776685.88563: getting the next task for host managed_node3 9704 1726776685.88568: done getting next task for host managed_node3 9704 1726776685.88569: ^ task is: TASK: meta (flush_handlers) 9704 1726776685.88570: ^ state is: HOST STATE: block=3, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776685.88574: getting variables 9704 1726776685.88575: in VariableManager get_vars() 9704 1726776685.88594: Calling all_inventory to load vars for managed_node3 9704 1726776685.88596: Calling groups_inventory to load vars for managed_node3 9704 1726776685.88597: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776685.88604: Calling all_plugins_play to load vars for managed_node3 9704 1726776685.88606: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776685.88607: Calling groups_plugins_play to load vars for managed_node3 9704 1726776685.88704: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776685.88812: done with get_vars() 9704 1726776685.88820: done getting variables 9704 1726776685.88873: in VariableManager get_vars() 9704 1726776685.88883: Calling all_inventory to load vars for managed_node3 9704 1726776685.88885: Calling groups_inventory to load vars for managed_node3 9704 1726776685.88886: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776685.88889: Calling all_plugins_play to load vars for managed_node3 9704 1726776685.88890: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776685.88892: Calling groups_plugins_play to load vars for managed_node3 9704 1726776685.88990: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776685.89098: done with get_vars() 9704 1726776685.89108: done queuing things up, now waiting for results queue to drain 9704 1726776685.89110: results queue empty 9704 1726776685.89110: checking for any_errors_fatal 9704 1726776685.89115: done checking for any_errors_fatal 9704 1726776685.89115: checking for max_fail_percentage 9704 1726776685.89116: done checking for max_fail_percentage 9704 1726776685.89116: checking to see if all hosts have failed and the running result is not ok 9704 1726776685.89116: done checking to see if all hosts have failed 9704 1726776685.89117: getting the remaining hosts for this loop 9704 1726776685.89117: done getting the remaining hosts for this loop 9704 1726776685.89119: getting the next task for host managed_node3 9704 1726776685.89122: done getting next task for host managed_node3 9704 1726776685.89123: ^ task is: TASK: meta (flush_handlers) 9704 1726776685.89124: ^ state is: HOST STATE: block=4, task=1, rescue=0, always=0, handlers=0, run_state=1, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776685.89126: getting variables 9704 1726776685.89127: in VariableManager get_vars() 9704 1726776685.89136: Calling all_inventory to load vars for managed_node3 9704 1726776685.89137: Calling groups_inventory to load vars for managed_node3 9704 1726776685.89138: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776685.89142: Calling all_plugins_play to load vars for managed_node3 9704 1726776685.89143: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776685.89144: Calling groups_plugins_play to load vars for managed_node3 9704 1726776685.89218: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776685.89319: done with get_vars() 9704 1726776685.89325: done getting variables 9704 1726776685.89354: in VariableManager get_vars() 9704 1726776685.89363: Calling all_inventory to load vars for managed_node3 9704 1726776685.89365: Calling groups_inventory to load vars for managed_node3 9704 1726776685.89366: Calling all_plugins_inventory to load vars for managed_node3 9704 1726776685.89369: Calling all_plugins_play to load vars for managed_node3 9704 1726776685.89370: Calling groups_plugins_inventory to load vars for managed_node3 9704 1726776685.89372: Calling groups_plugins_play to load vars for managed_node3 9704 1726776685.89446: '/usr/local/lib/python3.12/site-packages/ansible/plugins/connection/__init__' skipped due to reserved name 9704 1726776685.89548: done with get_vars() 9704 1726776685.89556: done queuing things up, now waiting for results queue to drain 9704 1726776685.89559: results queue empty 9704 1726776685.89560: checking for any_errors_fatal 9704 1726776685.89561: done checking for any_errors_fatal 9704 1726776685.89562: checking for max_fail_percentage 9704 1726776685.89562: done checking for max_fail_percentage 9704 1726776685.89562: checking to see if all hosts have failed and the running result is not ok 9704 1726776685.89563: done checking to see if all hosts have failed 9704 1726776685.89563: getting the remaining hosts for this loop 9704 1726776685.89563: done getting the remaining hosts for this loop 9704 1726776685.89565: getting the next task for host managed_node3 9704 1726776685.89567: done getting next task for host managed_node3 9704 1726776685.89567: ^ task is: None 9704 1726776685.89568: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False 9704 1726776685.89568: done queuing things up, now waiting for results queue to drain 9704 1726776685.89569: results queue empty 9704 1726776685.89569: checking for any_errors_fatal 9704 1726776685.89569: done checking for any_errors_fatal 9704 1726776685.89570: checking for max_fail_percentage 9704 1726776685.89570: done checking for max_fail_percentage 9704 1726776685.89571: checking to see if all hosts have failed and the running result is not ok 9704 1726776685.89571: done checking to see if all hosts have failed 9704 1726776685.89572: getting the next task for host managed_node3 9704 1726776685.89573: done getting next task for host managed_node3 9704 1726776685.89574: ^ task is: None 9704 1726776685.89574: ^ state is: HOST STATE: block=5, task=0, rescue=0, always=0, handlers=0, run_state=5, fail_state=0, pre_flushing_run_state=1, update_handlers=True, pending_setup=False, tasks child state? (None), rescue child state? (None), always child state? (None), did rescue? False, did start at task? False PLAY RECAP ********************************************************************* managed_node3 : ok=84 changed=12 unreachable=0 failed=0 skipped=46 rescued=0 ignored=0 Thursday 19 September 2024 16:11:25 -0400 (0:00:00.487) 0:00:38.825 **** =============================================================================== fedora.linux_system_roles.kernel_settings : Ensure required packages are installed --- 2.92s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:12 Ensure python command exists for tests below ---------------------------- 2.90s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:14 fedora.linux_system_roles.kernel_settings : Ensure required packages are installed --- 2.85s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:12 fedora.linux_system_roles.kernel_settings : Ensure required packages are installed --- 2.83s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:12 fedora.linux_system_roles.kernel_settings : Tuned apply settings -------- 1.55s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:157 Gathering Facts --------------------------------------------------------- 1.18s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tests_simple_settings.yml:2 fedora.linux_system_roles.kernel_settings : Ensure kernel_settings is in active_profile --- 0.76s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:91 fedora.linux_system_roles.kernel_settings : Restart tuned to apply active profile, mode changes --- 0.74s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:149 fedora.linux_system_roles.kernel_settings : Ensure required services are enabled and started --- 0.69s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:67 fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory --- 0.68s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:50 fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory --- 0.68s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:50 fedora.linux_system_roles.kernel_settings : Apply kernel settings ------- 0.68s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:112 fedora.linux_system_roles.kernel_settings : Apply kernel settings ------- 0.67s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:112 fedora.linux_system_roles.kernel_settings : Find tuned profile parent directory --- 0.67s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:50 fedora.linux_system_roles.kernel_settings : Apply kernel settings ------- 0.66s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:112 Put expected contents into temporary file ------------------------------- 0.65s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/assert_kernel_settings_conf_files.yml:8 Ensure kernel_settings is not in active_profile ------------------------- 0.65s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:46 Set profile_mode to auto ------------------------------------------------ 0.63s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/tests/kernel_settings/tasks/cleanup.yml:57 fedora.linux_system_roles.kernel_settings : Set profile_mode to manual --- 0.62s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/main.yml:99 fedora.linux_system_roles.kernel_settings : Check that settings are applied correctly --- 0.61s /tmp/collections-uMf/ansible_collections/fedora/linux_system_roles/roles/kernel_settings/tasks/verify_settings.yml:2 9704 1726776685.89654: RUNNING CLEANUP