ansible-playbook 2.9.27
  config file = /etc/ansible/ansible.cfg
  configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
  ansible python module location = /usr/lib/python2.7/site-packages/ansible
  executable location = /usr/bin/ansible-playbook
  python version = 2.7.5 (default, Nov 14 2023, 16:14:06) [GCC 4.8.5 20150623 (Red Hat 4.8.5-44)]
Using /etc/ansible/ansible.cfg as config file
[WARNING]: running playbook inside collection fedora.linux_system_roles
Skipping callback 'actionable', as we already have a stdout callback.
Skipping callback 'counter_enabled', as we already have a stdout callback.
Skipping callback 'debug', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'dense', as we already have a stdout callback.
Skipping callback 'full_skip', as we already have a stdout callback.
Skipping callback 'json', as we already have a stdout callback.
Skipping callback 'minimal', as we already have a stdout callback.
Skipping callback 'null', as we already have a stdout callback.
Skipping callback 'oneline', as we already have a stdout callback.
Skipping callback 'selective', as we already have a stdout callback.
Skipping callback 'skippy', as we already have a stdout callback.
Skipping callback 'stderr', as we already have a stdout callback.
Skipping callback 'unixy', as we already have a stdout callback.
Skipping callback 'yaml', as we already have a stdout callback.

PLAYBOOK: tests_mesh_cert.yml **************************************************
1 plays in /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml

PLAY [Test mesh tunnels with certificates] *************************************

TASK [Gathering Facts] *********************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:3
Sunday 13 October 2024  03:54:43 -0400 (0:00:00.037)       0:00:00.037 ******** 
ok: [managed-node2]
META: ran handlers

TASK [Set up test environment] *************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:16
Sunday 13 October 2024  03:54:44 -0400 (0:00:01.238)       0:00:01.276 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml for managed-node2

TASK [Set platform/version specific variables] *********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:2
Sunday 13 October 2024  03:54:44 -0400 (0:00:00.069)       0:00:01.345 ******** 

TASK [fedora.linux_system_roles.vpn : Ensure ansible_facts used by role] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:2
Sunday 13 October 2024  03:54:44 -0400 (0:00:00.178)       0:00:01.524 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Check if system is ostree] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:10
Sunday 13 October 2024  03:54:45 -0400 (0:00:00.066)       0:00:01.591 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.vpn : Set flag to indicate system is ostree] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:15
Sunday 13 October 2024  03:54:45 -0400 (0:00:00.761)       0:00:02.352 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__vpn_is_ostree": false
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Set platform/version specific variables] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:19
Sunday 13 October 2024  03:54:45 -0400 (0:00:00.180)       0:00:02.532 ******** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__vpn_nss_location": "/etc/ipsec.d"
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [Change inventory_hostname] ***********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:11
Sunday 13 October 2024  03:54:46 -0400 (0:00:00.149)       0:00:02.682 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Add sample main host] ****************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:18
Sunday 13 October 2024  03:54:46 -0400 (0:00:00.067)       0:00:02.749 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Create nss directory for testing] ****************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:22
Sunday 13 October 2024  03:54:46 -0400 (0:00:00.057)       0:00:02.807 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "path": "/etc/ipsec.d", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 4096, 
    "state": "directory", 
    "uid": 0
}

TASK [Create /etc/ipsec.d directory for testing] *******************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:28
Sunday 13 October 2024  03:54:46 -0400 (0:00:00.531)       0:00:03.338 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "path": "/etc/ipsec.d", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 4096, 
    "state": "directory", 
    "uid": 0
}

TASK [Dynamically add more hosts] **********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:34
Sunday 13 October 2024  03:54:47 -0400 (0:00:00.291)       0:00:03.629 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/add_hosts.yml for managed-node2

TASK [Dynamically add multiple hosts for testing] ******************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/add_hosts.yml:2
Sunday 13 October 2024  03:54:47 -0400 (0:00:00.025)       0:00:03.655 ******** 
creating host via 'add_host': hostname=host01.local
changed: [managed-node2] => (item=1) => {
    "add_host": {
        "groups": [
            "testing"
        ], 
        "host_name": "host01.local", 
        "host_vars": {
            "cert_name": "dyn_cert", 
            "current_ip": "169.254.1.1", 
            "current_subnet": "169.254.0.0/16"
        }
    }, 
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": 1
}
creating host via 'add_host': hostname=host02.local
changed: [managed-node2] => (item=2) => {
    "add_host": {
        "groups": [
            "testing"
        ], 
        "host_name": "host02.local", 
        "host_vars": {
            "cert_name": "dyn_cert", 
            "current_ip": "169.254.1.1", 
            "current_subnet": "169.254.0.0/16"
        }
    }, 
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": 2
}

TASK [Create mock vpn_connections] *********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/add_hosts.yml:11
Sunday 13 October 2024  03:54:47 -0400 (0:00:00.062)       0:00:03.717 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "vpn_connections": [
            {
                "hosts": {
                    "host01.local": "", 
                    "host02.local": "", 
                    "managed-node1": "", 
                    "managed-node2": "", 
                    "managed-node3": ""
                }
            }
        ]
    }, 
    "changed": false
}

TASK [Create sample policies directory for testing] ****************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:40
Sunday 13 October 2024  03:54:47 -0400 (0:00:00.111)       0:00:03.829 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "path": "/etc/ipsec.d/policies", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 4096, 
    "state": "directory", 
    "uid": 0
}

TASK [Create sample policy files for testing] **********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:46
Sunday 13 October 2024  03:54:47 -0400 (0:00:00.414)       0:00:04.244 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/etc/ipsec.d/policies/private", 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Create sample policy files for testing] **********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:52
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.339)       0:00:04.584 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/etc/ipsec.d/policies/private-or-clear", 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Create sample policy files for testing] **********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:58
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.355)       0:00:04.939 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "dest": "/etc/ipsec.d/policies/clear", 
    "gid": 0, 
    "group": "root", 
    "mode": "0600", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 0, 
    "state": "file", 
    "uid": 0
}

TASK [Ensure firewalld not running if not testing firewall] ********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/setup_test.yml:67
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.365)       0:00:05.304 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Add extra options to check] **********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:21
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.061)       0:00:05.366 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "vpn_connections": [
            {
                "auth_method": "cert", 
                "hosts": {
                    "host01.local": "", 
                    "host02.local": "", 
                    "managed-node1": "", 
                    "managed-node2": "", 
                    "managed-node3": ""
                }, 
                "opportunistic": true, 
                "policies": [
                    {
                        "cidr": "203.0.113.0/24", 
                        "policy": "private"
                    }, 
                    {
                        "cidr": "198.51.100.0/24", 
                        "policy": "clear"
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [Add cert_name fact to controller] ****************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:36
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.071)       0:00:05.437 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "cert_name": "main_cert"
    }, 
    "changed": false
}

TASK [Use vpn role] ************************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:40
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.048)       0:00:05.486 ******** 

TASK [fedora.linux_system_roles.vpn : Set platform/version specific variables] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:3
Sunday 13 October 2024  03:54:48 -0400 (0:00:00.068)       0:00:05.554 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml for managed-node2

TASK [fedora.linux_system_roles.vpn : Ensure ansible_facts used by role] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:2
Sunday 13 October 2024  03:54:49 -0400 (0:00:00.036)       0:00:05.590 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Check if system is ostree] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:10
Sunday 13 October 2024  03:54:49 -0400 (0:00:00.062)       0:00:05.653 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Set flag to indicate system is ostree] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:15
Sunday 13 October 2024  03:54:49 -0400 (0:00:00.058)       0:00:05.712 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Set platform/version specific variables] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:19
Sunday 13 October 2024  03:54:49 -0400 (0:00:00.050)       0:00:05.762 ******** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__vpn_nss_location": "/etc/ipsec.d"
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Ensure required packages are installed] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:6
Sunday 13 October 2024  03:54:49 -0400 (0:00:00.081)       0:00:05.844 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "changes": {
        "installed": [
            "libreswan"
        ]
    }, 
    "rc": 0, 
    "results": [
        "iproute-4.11.0-30.el7.x86_64 providing iproute is already installed", 
        "Loaded plugins: fastestmirror\nLoading mirror speeds from cached hostfile\n * epel: d2lzkl7pfhq30w.cloudfront.net\n * epel-debuginfo: d2lzkl7pfhq30w.cloudfront.net\n * epel-source: d2lzkl7pfhq30w.cloudfront.net\nResolving Dependencies\n--> Running transaction check\n---> Package libreswan.x86_64 0:3.25-9.1.el7_8 will be installed\n--> Processing Dependency: unbound-libs >= 1.6.6 for package: libreswan-3.25-9.1.el7_8.x86_64\n--> Processing Dependency: libunbound.so.2()(64bit) for package: libreswan-3.25-9.1.el7_8.x86_64\n--> Processing Dependency: libldns.so.1()(64bit) for package: libreswan-3.25-9.1.el7_8.x86_64\n--> Running transaction check\n---> Package ldns.x86_64 0:1.6.16-10.el7 will be installed\n--> Processing Dependency: libpcap.so.1()(64bit) for package: ldns-1.6.16-10.el7.x86_64\n---> Package unbound-libs.x86_64 0:1.6.6-5.el7_8 will be installed\n--> Running transaction check\n---> Package libpcap.x86_64 14:1.5.3-13.el7_9 will be installed\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package            Arch         Version                    Repository     Size\n================================================================================\nInstalling:\n libreswan          x86_64       3.25-9.1.el7_8             updates       1.4 M\nInstalling for dependencies:\n ldns               x86_64       1.6.16-10.el7              base          476 k\n libpcap            x86_64       14:1.5.3-13.el7_9          updates       139 k\n unbound-libs       x86_64       1.6.6-5.el7_8              base          406 k\n\nTransaction Summary\n================================================================================\nInstall  1 Package (+3 Dependent packages)\n\nTotal download size: 2.4 M\nInstalled size: 7.5 M\nDownloading packages:\n--------------------------------------------------------------------------------\nTotal                                              5.3 MB/s | 2.4 MB  00:00     \nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Installing : 14:libpcap-1.5.3-13.el7_9.x86_64                             1/4 \n  Installing : ldns-1.6.16-10.el7.x86_64                                    2/4 \n  Installing : unbound-libs-1.6.6-5.el7_8.x86_64                            3/4 \n  Installing : libreswan-3.25-9.1.el7_8.x86_64                              4/4 \nwarning: /etc/ipsec.d/policies/clear created as /etc/ipsec.d/policies/clear.rpmnew\nwarning: /etc/ipsec.d/policies/private created as /etc/ipsec.d/policies/private.rpmnew\nwarning: /etc/ipsec.d/policies/private-or-clear created as /etc/ipsec.d/policies/private-or-clear.rpmnew\n  Verifying  : ldns-1.6.16-10.el7.x86_64                                    1/4 \n  Verifying  : libreswan-3.25-9.1.el7_8.x86_64                              2/4 \n  Verifying  : unbound-libs-1.6.6-5.el7_8.x86_64                            3/4 \n  Verifying  : 14:libpcap-1.5.3-13.el7_9.x86_64                             4/4 \n\nInstalled:\n  libreswan.x86_64 0:3.25-9.1.el7_8                                             \n\nDependency Installed:\n  ldns.x86_64 0:1.6.16-10.el7              libpcap.x86_64 14:1.5.3-13.el7_9     \n  unbound-libs.x86_64 0:1.6.6-5.el7_8     \n\nComplete!\n"
    ]
}
lsrpackages: iproute libreswan

TASK [fedora.linux_system_roles.vpn : Configure firewall] **********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:13
Sunday 13 October 2024  03:54:56 -0400 (0:00:07.000)       0:00:12.844 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/firewall.yml for managed-node2

TASK [Ensure the vpn ports status with the firewall role] **********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/firewall.yml:3
Sunday 13 October 2024  03:54:56 -0400 (0:00:00.037)       0:00:12.882 ******** 

TASK [fedora.linux_system_roles.firewall : Setup firewalld] ********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:2
Sunday 13 October 2024  03:54:56 -0400 (0:00:00.100)       0:00:12.983 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml for managed-node2

TASK [fedora.linux_system_roles.firewall : Ensure ansible_facts used by role] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:2
Sunday 13 October 2024  03:54:56 -0400 (0:00:00.067)       0:00:13.050 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Check if system is ostree] **********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:10
Sunday 13 October 2024  03:54:56 -0400 (0:00:00.055)       0:00:13.106 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.firewall : Set flag to indicate system is ostree] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:15
Sunday 13 October 2024  03:54:56 -0400 (0:00:00.471)       0:00:13.578 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__firewall_is_ostree": false
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.firewall : Check if transactional-update exists in /sbin] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:22
Sunday 13 October 2024  03:54:57 -0400 (0:00:00.088)       0:00:13.666 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "exists": false
    }
}

TASK [fedora.linux_system_roles.firewall : Set flag if transactional-update exists] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:27
Sunday 13 October 2024  03:54:57 -0400 (0:00:00.313)       0:00:13.980 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__firewall_is_transactional": false
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.firewall : Install firewalld] ******************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31
Sunday 13 October 2024  03:54:57 -0400 (0:00:00.056)       0:00:14.036 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "rc": 0, 
    "results": [
        "firewalld-0.6.3-13.el7_9.noarch providing firewalld is already installed"
    ]
}
lsrpackages: firewalld

TASK [fedora.linux_system_roles.firewall : Notify user that reboot is needed to apply changes] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:43
Sunday 13 October 2024  03:54:58 -0400 (0:00:00.585)       0:00:14.622 ******** 
skipping: [managed-node2] => {}

TASK [fedora.linux_system_roles.firewall : Reboot transactional update systems] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:48
Sunday 13 October 2024  03:54:58 -0400 (0:00:00.055)       0:00:14.678 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Fail if reboot is needed and not set] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:53
Sunday 13 October 2024  03:54:58 -0400 (0:00:00.078)       0:00:14.756 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Collect service facts] **************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:5
Sunday 13 October 2024  03:54:58 -0400 (0:00:00.095)       0:00:14.852 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Attempt to stop and disable conflicting services] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:9
Sunday 13 October 2024  03:54:58 -0400 (0:00:00.077)       0:00:14.930 ******** 
skipping: [managed-node2] => (item=nftables)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "nftables", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=iptables)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "iptables", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=ufw)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "ufw", 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Unmask firewalld service] ***********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22
Sunday 13 October 2024  03:54:58 -0400 (0:00:00.100)       0:00:15.031 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "name": "firewalld", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "system.slice basic.target polkit.service dbus.service", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "network-pre.target shutdown.target", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "BusName": "org.fedoraproject.FirewallD1", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "iptables.service ip6tables.service ebtables.service shutdown.target ipset.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "firewalld - dynamic firewall daemon", 
        "DevicePolicy": "auto", 
        "Documentation": "man:firewalld(1)", 
        "EnvironmentFile": "/etc/sysconfig/firewalld (ignore_errors=yes)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/firewalld.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "firewalld.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "mixed", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "firewalld.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Requires": "basic.target system.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "null", 
        "StandardInput": "null", 
        "StandardOutput": "null", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "dbus", 
        "UMask": "0022", 
        "UnitFilePreset": "enabled", 
        "UnitFileState": "disabled", 
        "Wants": "network-pre.target", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.firewall : Enable and start firewalld service] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28
Sunday 13 October 2024  03:54:59 -0400 (0:00:00.802)       0:00:15.833 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "enabled": true, 
    "name": "firewalld", 
    "state": "started", 
    "status": {
        "ActiveEnterTimestampMonotonic": "0", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "inactive", 
        "After": "system.slice basic.target polkit.service dbus.service", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "no", 
        "AssertTimestampMonotonic": "0", 
        "Before": "network-pre.target shutdown.target", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "BusName": "org.fedoraproject.FirewallD1", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "no", 
        "ConditionTimestampMonotonic": "0", 
        "Conflicts": "ip6tables.service ebtables.service shutdown.target ipset.service iptables.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "firewalld - dynamic firewall daemon", 
        "DevicePolicy": "auto", 
        "Documentation": "man:firewalld(1)", 
        "EnvironmentFile": "/etc/sysconfig/firewalld (ignore_errors=yes)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "0", 
        "ExecMainStartTimestampMonotonic": "0", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/firewalld.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "firewalld.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestampMonotonic": "0", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "mixed", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "0", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "firewalld.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Requires": "basic.target system.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "null", 
        "StandardInput": "null", 
        "StandardOutput": "null", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "dead", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "dbus", 
        "UMask": "0022", 
        "UnitFilePreset": "enabled", 
        "UnitFileState": "disabled", 
        "Wants": "network-pre.target", 
        "WatchdogTimestampMonotonic": "0", 
        "WatchdogUSec": "0"
    }
}

TASK [fedora.linux_system_roles.firewall : Check if previous replaced is defined] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:34
Sunday 13 October 2024  03:55:00 -0400 (0:00:00.858)       0:00:16.691 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__firewall_previous_replaced": false, 
        "__firewall_python_cmd": "/usr/bin/python", 
        "__firewall_report_changed": true
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.firewall : Get config files, checksums before and remove] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:43
Sunday 13 October 2024  03:55:00 -0400 (0:00:00.080)       0:00:16.772 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Tell firewall module it is able to report changed] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:55
Sunday 13 October 2024  03:55:00 -0400 (0:00:00.101)       0:00:16.873 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Configure firewall] *****************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71
Sunday 13 October 2024  03:55:00 -0400 (0:00:00.065)       0:00:16.939 ******** 
changed: [managed-node2] => (item={u'state': u'enabled', u'service': u'ipsec'}) => {
    "__firewall_changed": true, 
    "ansible_loop_var": "item", 
    "changed": true, 
    "item": {
        "service": "ipsec", 
        "state": "enabled"
    }
}

TASK [fedora.linux_system_roles.firewall : Gather firewall config information] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:120
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.961)       0:00:17.900 ******** 
skipping: [managed-node2] => (item={u'state': u'enabled', u'service': u'ipsec'})  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": {
        "service": "ipsec", 
        "state": "enabled"
    }, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:130
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.086)       0:00:17.987 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Gather firewall config if no arguments] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:139
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.066)       0:00:18.054 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Update firewalld_config fact] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:144
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.067)       0:00:18.122 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Get config files, checksums after] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:153
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.052)       0:00:18.174 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Calculate what has changed] *********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:163
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.068)       0:00:18.242 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.firewall : Show diffs] *************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:169
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.076)       0:00:18.319 ******** 
skipping: [managed-node2] => {}

TASK [fedora.linux_system_roles.vpn : Configure selinux] ***********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:16
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.054)       0:00:18.374 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml for managed-node2

TASK [fedora.linux_system_roles.vpn : Populate service facts] ******************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml:3
Sunday 13 October 2024  03:55:01 -0400 (0:00:00.049)       0:00:18.423 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.fedoraproject.FirewallD1.service": {
                "name": "dbus-org.fedoraproject.FirewallD1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "ipsec.service": {
                "name": "ipsec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "unbound-anchor.service": {
                "name": "unbound-anchor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Get the ipsec tcp service ports] *********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml:12
Sunday 13 October 2024  03:55:02 -0400 (0:00:01.137)       0:00:19.561 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Initialize _vpn_selinux] *****************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml:20
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.090)       0:00:19.651 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Add the ipsec service ports to _vpn_selinux] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml:24
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.084)       0:00:19.736 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Ensure the service and the ports status with the selinux role] ***********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml:31
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.083)       0:00:19.820 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Ensure OpenSSL is installed on the controller] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:26
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.097)       0:00:19.918 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Enforce default auth method as needed] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:34
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.070)       0:00:19.988 ******** 
ok: [managed-node2 -> localhost] => {
    "ansible_facts": {
        "vpn_connections": [
            {
                "auth_method": "cert", 
                "hosts": {
                    "host01.local": "", 
                    "host02.local": "", 
                    "managed-node1": "", 
                    "managed-node2": "", 
                    "managed-node3": ""
                }, 
                "opportunistic": true, 
                "policies": [
                    {
                        "cidr": "203.0.113.0/24", 
                        "policy": "private"
                    }, 
                    {
                        "cidr": "198.51.100.0/24", 
                        "policy": "clear"
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Make sure that the hosts list is not empty] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:44
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.086)       0:00:20.075 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Make sure there is at least one pair of hosts in each connection] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:59
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.089)       0:00:20.164 ******** 
ok: [managed-node2 -> localhost] => {
    "ansible_facts": {
        "vpn_connections": [
            {
                "auth_method": "cert", 
                "hosts": {
                    "host01.local": "", 
                    "host02.local": "", 
                    "managed-node1": "", 
                    "managed-node2": "", 
                    "managed-node3": ""
                }, 
                "opportunistic": true, 
                "policies": [
                    {
                        "cidr": "203.0.113.0/24", 
                        "policy": "private"
                    }, 
                    {
                        "cidr": "198.51.100.0/24", 
                        "policy": "clear"
                    }
                ]
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Ensure cert_names are populated when auth_method is cert] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:73
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.078)       0:00:20.242 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Generate PSKs or use provided shared_key_content] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:100
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.186)       0:00:20.429 ******** 
ok: [managed-node2 -> localhost] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Set psks for hosts] **********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:137
Sunday 13 October 2024  03:55:03 -0400 (0:00:00.100)       0:00:20.530 ******** 
ok: [managed-node2] => {
    "censored": "the output has been hidden due to the fact that 'no_log: true' was specified for this result", 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Create ipsec.conf files] *****************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:161
Sunday 13 October 2024  03:55:04 -0400 (0:00:00.188)       0:00:20.719 ******** 

TASK [fedora.linux_system_roles.vpn : Check if secrets file already exists] ****
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:169
Sunday 13 October 2024  03:55:04 -0400 (0:00:00.079)       0:00:20.798 ******** 

TASK [fedora.linux_system_roles.vpn : Create ipsec.secrets files] **************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:175
Sunday 13 October 2024  03:55:04 -0400 (0:00:00.070)       0:00:20.868 ******** 

TASK [fedora.linux_system_roles.vpn : Build opportunistic configuration] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:185
Sunday 13 October 2024  03:55:04 -0400 (0:00:00.066)       0:00:20.935 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml for managed-node2

TASK [fedora.linux_system_roles.vpn : Set current IP fact for each host] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:3
Sunday 13 October 2024  03:55:04 -0400 (0:00:00.143)       0:00:21.079 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "current_ip": "10.31.40.241"
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Set IP with prefix register] *************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:7
Sunday 13 October 2024  03:55:04 -0400 (0:00:00.130)       0:00:21.209 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": "set -euo pipefail\nip addr show | grep 10.31.40.241 | awk '{print $2}'", 
    "delta": "0:00:00.005975", 
    "end": "2024-10-13 03:55:05.066298", 
    "rc": 0, 
    "start": "2024-10-13 03:55:05.060323"
}

STDOUT:

10.31.40.241/22

TASK [fedora.linux_system_roles.vpn : Set net CIDR fact] ***********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:14
Sunday 13 October 2024  03:55:05 -0400 (0:00:00.503)       0:00:21.712 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "current_subnet": "10.31.40.0/22"
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Set policies fact] ***********************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:18
Sunday 13 October 2024  03:55:05 -0400 (0:00:00.064)       0:00:21.776 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "policies": [
            {
                "cidr": "203.0.113.0/24", 
                "policy": "private"
            }, 
            {
                "cidr": "198.51.100.0/24", 
                "policy": "clear"
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Apply the default policy as needed] ******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:22
Sunday 13 October 2024  03:55:05 -0400 (0:00:00.060)       0:00:21.837 ******** 
[WARNING]: The value 'False' is not a valid IP address or network, passing this
value to ipaddr filter might result in breaking change in future.
ok: [managed-node2 -> localhost] => {
    "ansible_facts": {
        "policies": [
            {
                "cidr": "203.0.113.0/24", 
                "policy": "private"
            }, 
            {
                "cidr": "198.51.100.0/24", 
                "policy": "clear"
            }, 
            {
                "cidr": "10.31.40.0/22", 
                "policy": "private-or-clear"
            }, 
            {
                "cidr": "169.254.0.0/16", 
                "policy": "private-or-clear"
            }
        ]
    }, 
    "changed": false
}

TASK [fedora.linux_system_roles.vpn : Write tunnel policies for each network] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:48
Sunday 13 October 2024  03:55:05 -0400 (0:00:00.223)       0:00:22.061 ******** 
changed: [managed-node2] => (item=clear) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "checksum": "b3813bc54754cc738fae6c9fd8011b7165c12da4", 
    "dest": "/etc/ipsec.d/policies/clear", 
    "gid": 0, 
    "group": "root", 
    "item": "clear", 
    "md5sum": "e8ed4c5e8408ffcdea43aa3810fd3442", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 57, 
    "src": "/root/.ansible/tmp/ansible-tmp-1728806105.58-10687-44290174375197/source", 
    "state": "file", 
    "uid": 0
}
changed: [managed-node2] => (item=private-or-clear) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "checksum": "4c64431a5e978157ad99acdd5526ffd654a4402d", 
    "dest": "/etc/ipsec.d/policies/private-or-clear", 
    "gid": 0, 
    "group": "root", 
    "item": "private-or-clear", 
    "md5sum": "0735d0a3ca6d3e8896f021faa6a24fd7", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 70, 
    "src": "/root/.ansible/tmp/ansible-tmp-1728806106.42-10687-124215217131903/source", 
    "state": "file", 
    "uid": 0
}
changed: [managed-node2] => (item=private) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "checksum": "a4789f2036166e565e40fdcea385e5ff82ad16b6", 
    "dest": "/etc/ipsec.d/policies/private", 
    "gid": 0, 
    "group": "root", 
    "item": "private", 
    "md5sum": "8a1d49feea7237a76e10628ddc628987", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "unconfined_u:object_r:etc_t:s0", 
    "size": 56, 
    "src": "/root/.ansible/tmp/ansible-tmp-1728806107.07-10687-235592500081143/source", 
    "state": "file", 
    "uid": 0
}
NOTIFIED HANDLER fedora.linux_system_roles.vpn : Restart vpn service and wait for ssh conn to return for managed-node2
NOTIFIED HANDLER fedora.linux_system_roles.vpn : Restart vpn service and wait for ssh conn to return for managed-node2
NOTIFIED HANDLER fedora.linux_system_roles.vpn : Send pings to initialize mesh connections for managed-node2
NOTIFIED HANDLER fedora.linux_system_roles.vpn : Send pings to initialize mesh connections for managed-node2

TASK [fedora.linux_system_roles.vpn : Deploy opportunistic configuration to each node] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:58
Sunday 13 October 2024  03:55:07 -0400 (0:00:02.253)       0:00:24.314 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "checksum": "433a116d6d92ed3a1793ba1a066c4f1436b58452", 
    "dest": "/etc/ipsec.d/mesh.conf", 
    "gid": 0, 
    "group": "root", 
    "md5sum": "755e302046e7ac0e477c61e07c01dfab", 
    "mode": "0644", 
    "owner": "root", 
    "secontext": "system_u:object_r:ipsec_key_file_t:s0", 
    "size": 859, 
    "src": "/root/.ansible/tmp/ansible-tmp-1728806107.81-10848-75047567506508/source", 
    "state": "file", 
    "uid": 0
}

RUNNING HANDLER [fedora.linux_system_roles.vpn : Restart vpn service and wait for ssh conn to return] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/handlers/main.yml:2
Sunday 13 October 2024  03:55:08 -0400 (0:00:00.797)       0:00:25.112 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml for managed-node2

RUNNING HANDLER [fedora.linux_system_roles.vpn : Enable and restart vpn services] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:2
Sunday 13 October 2024  03:55:08 -0400 (0:00:00.072)       0:00:25.184 ******** 
changed: [managed-node2] => (item=ipsec) => {
    "ansible_job_id": "897922292461.10771", 
    "ansible_loop_var": "item", 
    "changed": true, 
    "finished": 0, 
    "item": "ipsec", 
    "results_file": "/root/.ansible_async/897922292461.10771", 
    "started": 1
}

RUNNING HANDLER [fedora.linux_system_roles.vpn : Wait for ssh connection to return] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:11
Sunday 13 October 2024  03:55:09 -0400 (0:00:00.868)       0:00:26.053 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "elapsed": 3
}

RUNNING HANDLER [fedora.linux_system_roles.vpn : Send pings to initialize mesh connections] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/handlers/main.yml:6
Sunday 13 October 2024  03:55:13 -0400 (0:00:03.907)       0:00:29.960 ******** 
skipping: [managed-node2] => (item=managed-node2)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "managed-node2", 
    "skip_reason": "Conditional result was False"
}

RUNNING HANDLER [fedora.linux_system_roles.vpn : Restart vpn service and wait for ssh conn to return] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/handlers/main.yml:2
Sunday 13 October 2024  03:55:13 -0400 (0:00:00.060)       0:00:30.020 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml for managed-node2

RUNNING HANDLER [fedora.linux_system_roles.vpn : Enable and restart vpn services] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:2
Sunday 13 October 2024  03:55:13 -0400 (0:00:00.067)       0:00:30.088 ******** 
changed: [managed-node2] => (item=ipsec) => {
    "ansible_job_id": "581107889788.11493", 
    "ansible_loop_var": "item", 
    "changed": true, 
    "finished": 0, 
    "item": "ipsec", 
    "results_file": "/root/.ansible_async/581107889788.11493", 
    "started": 1
}

RUNNING HANDLER [fedora.linux_system_roles.vpn : Wait for ssh connection to return] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:11
Sunday 13 October 2024  03:55:14 -0400 (0:00:00.622)       0:00:30.710 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "elapsed": 3
}

RUNNING HANDLER [fedora.linux_system_roles.vpn : Send pings to initialize mesh connections] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/handlers/main.yml:6
Sunday 13 October 2024  03:55:17 -0400 (0:00:03.595)       0:00:34.306 ******** 
skipping: [managed-node2] => (item=managed-node2)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "managed-node2", 
    "skip_reason": "Conditional result was False"
}
META: ran handlers

TASK [Assert file existence] ***************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:49
Sunday 13 October 2024  03:55:17 -0400 (0:00:00.075)       0:00:34.382 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/assert_mesh_conf_exists.yml for managed-node2

TASK [Stat mesh.conf file path] ************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/assert_mesh_conf_exists.yml:2
Sunday 13 October 2024  03:55:17 -0400 (0:00:00.060)       0:00:34.442 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "stat": {
        "atime": 1728806109.6468747, 
        "attr_flags": "e", 
        "attributes": [
            "extents"
        ], 
        "block_size": 4096, 
        "blocks": 8, 
        "charset": "us-ascii", 
        "checksum": "433a116d6d92ed3a1793ba1a066c4f1436b58452", 
        "ctime": 1728806108.4788742, 
        "dev": 51713, 
        "device_type": 0, 
        "executable": false, 
        "exists": true, 
        "gid": 0, 
        "gr_name": "root", 
        "inode": 262454, 
        "isblk": false, 
        "ischr": false, 
        "isdir": false, 
        "isfifo": false, 
        "isgid": false, 
        "islnk": false, 
        "isreg": true, 
        "issock": false, 
        "isuid": false, 
        "mimetype": "text/plain", 
        "mode": "0644", 
        "mtime": 1728806108.157874, 
        "nlink": 1, 
        "path": "/etc/ipsec.d/mesh.conf", 
        "pw_name": "root", 
        "readable": true, 
        "rgrp": true, 
        "roth": true, 
        "rusr": true, 
        "size": 859, 
        "uid": 0, 
        "version": "18446744072465560330", 
        "wgrp": false, 
        "woth": false, 
        "writeable": true, 
        "wusr": true, 
        "xgrp": false, 
        "xoth": false, 
        "xusr": false
    }
}

TASK [Assert that mesh.conf exists] ********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/assert_mesh_conf_exists.yml:7
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.381)       0:00:34.824 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Reset success flag] ******************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:52
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.066)       0:00:34.890 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "__vpn_success": true
    }, 
    "changed": false
}

TASK [Get and store mesh.conf file] ********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:56
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.032)       0:00:34.923 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6dnBuCgpjb25uIHByaXZhdGUKICAgICAgdHlwZT10dW5uZWwKICAgICAgbGVmdD0lZGVmYXVsdHJvdXRlCiAgICAgIGxlZnRpZD0lZnJvbWNlcnQKICAgICAgcmlnaHRpZD0lZnJvbWNlcnQKICAgICAgcmlnaHRyc2FzaWdrZXk9JWNlcnQKICAgICAgcmlnaHRjYT0lc2FtZQogICAgICBsZWZ0Y2VydD1tYWluX2NlcnQKICAgICAgYXV0bz1yb3V0ZQogICAgICByaWdodD0lb3Bwb3J0dW5pc3RpY2dyb3VwCiAgICAgIG5lZ290aWF0aW9uc2h1bnQ9aG9sZAogICAgICBmYWlsdXJlc2h1bnQ9ZHJvcAogICAgICBpa2V2Mj1pbnNpc3QKICAgICAga2V5aW5ndHJpZXM9MQogICAgICByZXRyYW5zbWl0LXRpbWVvdXQ9MnMKCmNvbm4gcHJpdmF0ZS1vci1jbGVhcgogICAgICB0eXBlPXR1bm5lbAogICAgICBsZWZ0PSVkZWZhdWx0cm91dGUKICAgICAgbGVmdGlkPSVmcm9tY2VydAogICAgICByaWdodGlkPSVmcm9tY2VydAogICAgICByaWdodHJzYXNpZ2tleT0lY2VydAogICAgICByaWdodGNhPSVzYW1lCiAgICAgIGxlZnRjZXJ0PW1haW5fY2VydAogICAgICBhdXRvPXJvdXRlCiAgICAgIHJpZ2h0PSVvcHBvcnR1bmlzdGljZ3JvdXAKICAgICAgbmVnb3RpYXRpb25zaHVudD1ob2xkCiAgICAgIGZhaWx1cmVzaHVudD1wYXNzdGhyb3VnaAogICAgICBpa2V2Mj1pbnNpc3QKICAgICAga2V5aW5ndHJpZXM9MQogICAgICByZXRyYW5zbWl0LXRpbWVvdXQ9MnMKCmNvbm4gY2xlYXIKICAgICAgdHlwZT1wYXNzdGhyb3VnaAogICAgICBsZWZ0PSVkZWZhdWx0cm91dGUKICAgICAgcmlnaHQ9JWdyb3VwCiAgICAgIGF1dG89cm91dGUKICAgICAgYXV0aGJ5PW5ldmVyCg==", 
    "encoding": "base64", 
    "source": "/etc/ipsec.d/mesh.conf"
}

TASK [Extract content for conn private] ****************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:61
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.408)       0:00:35.331 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "conn_private": "conn private\n      type=tunnel\n      left=%defaultroute\n      leftid=%fromcert\n      rightid=%fromcert\n      rightrsasigkey=%cert\n      rightca=%same\n      leftcert=main_cert\n      auto=route\n      right=%opportunisticgroup\n      negotiationshunt=hold\n      failureshunt=drop\n      ikev2=insist\n      keyingtries=1\n      retransmit-timeout=2s"
    }, 
    "changed": false
}

TASK [Extract content for conn private-or-clear] *******************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:65
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.075)       0:00:35.406 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "conn_private_or_clear": "conn private-or-clear\n      type=tunnel\n      left=%defaultroute\n      leftid=%fromcert\n      rightid=%fromcert\n      rightrsasigkey=%cert\n      rightca=%same\n      leftcert=main_cert\n      auto=route\n      right=%opportunisticgroup\n      negotiationshunt=hold\n      failureshunt=passthrough\n      ikev2=insist\n      keyingtries=1\n      retransmit-timeout=2s"
    }, 
    "changed": false
}

TASK [Extract content for conn clear] ******************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:69
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.068)       0:00:35.474 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "conn_clear": "conn clear\n      type=passthrough\n      left=%defaultroute\n      right=%group\n      auto=route\n      authby=never\n"
    }, 
    "changed": false
}

TASK [Check that private section contains correct information] *****************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:73
Sunday 13 October 2024  03:55:18 -0400 (0:00:00.062)       0:00:35.537 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Assert success for private part of conf file] ****************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:92
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.061)       0:00:35.598 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check that private-or-clear section contains correct information] ********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:97
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.068)       0:00:35.666 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Assert success for private or clear part of conf file] *******************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:116
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.080)       0:00:35.747 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check that clear section contains correct information] *******************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:121
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.081)       0:00:35.828 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Assert success for clear part of conf file] ******************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:131
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.073)       0:00:35.902 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get and store policies private file] *************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:136
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.070)       0:00:35.972 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6dnBuCgoyMDMuMC4xMTMuMC8yNAo=", 
    "encoding": "base64", 
    "source": "/etc/ipsec.d/policies/private"
}

TASK [Get and store policies private or clear file] ****************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:141
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.298)       0:00:36.271 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6dnBuCgoxMC4zMS40MC4wLzIyCjE2OS4yNTQuMC4wLzE2Cg==", 
    "encoding": "base64", 
    "source": "/etc/ipsec.d/policies/private-or-clear"
}

TASK [Get and store policies clear file] ***************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:146
Sunday 13 October 2024  03:55:19 -0400 (0:00:00.292)       0:00:36.563 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "content": "IwojIEFuc2libGUgbWFuYWdlZAojCiMgc3lzdGVtX3JvbGU6dnBuCgoxOTguNTEuMTAwLjAvMjQK", 
    "encoding": "base64", 
    "source": "/etc/ipsec.d/policies/clear"
}

TASK [Check contents of policy private file] ***********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:151
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.324)       0:00:36.887 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Assert success for policy private file] **********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:158
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.078)       0:00:36.966 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check contents of policy private or clear file] **************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:163
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.097)       0:00:37.064 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Assert success for policy private or clear file] *************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:172
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.080)       0:00:37.144 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check contents of policy clear file] *************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:177
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.066)       0:00:37.211 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Assert success for policy clear file] ************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:184
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.060)       0:00:37.271 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Check the firewall and the selinux port status] **************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:189
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.083)       0:00:37.355 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml for managed-node2

TASK [Get firewall service] ****************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml:9
Sunday 13 October 2024  03:55:20 -0400 (0:00:00.114)       0:00:37.469 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": [
        "firewall-cmd", 
        "--list-services"
    ], 
    "delta": "0:00:00.273307", 
    "end": "2024-10-13 03:55:21.484179", 
    "rc": 0, 
    "start": "2024-10-13 03:55:21.210872"
}

STDOUT:

dhcpv6-client ipsec ssh

TASK [Ensure 'ipsec' is in the firewalld service list] *************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml:14
Sunday 13 October 2024  03:55:21 -0400 (0:00:00.672)       0:00:38.141 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get associated selinux ports] ********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml:19
Sunday 13 October 2024  03:55:21 -0400 (0:00:00.094)       0:00:38.236 ******** 
ok: [managed-node2] => {
    "changed": false, 
    "cmd": "set -euo pipefail\nfirewall-cmd --info-service=ipsec |  egrep \"^ +ports: +\" | sed -e \"s/ *ports: //\"", 
    "delta": "0:00:00.249895", 
    "end": "2024-10-13 03:55:22.185719", 
    "rc": 0, 
    "start": "2024-10-13 03:55:21.935824"
}

STDOUT:

500/udp 4500/udp

TASK [Check associated selinux ports when vpn_manage_selinux is true] **********
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml:27
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.582)       0:00:38.818 ******** 
skipping: [managed-node2] => (item=500/udp)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "500/udp", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=4500/udp)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "4500/udp", 
    "skip_reason": "Conditional result was False"
}

TASK [Check for ansible_managed, fingerprint in generated files] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:192
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.101)       0:00:38.920 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml for managed-node2
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml for managed-node2
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml for managed-node2
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml for managed-node2

TASK [Get file] ****************************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:3
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.110)       0:00:39.031 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check for presence of ansible managed header, fingerprint] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:9
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.049)       0:00:39.080 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get file] ****************************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:3
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.103)       0:00:39.184 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check for presence of ansible managed header, fingerprint] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:9
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.049)       0:00:39.233 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get file] ****************************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:3
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.103)       0:00:39.336 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check for presence of ansible managed header, fingerprint] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:9
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.049)       0:00:39.386 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Get file] ****************************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:3
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.102)       0:00:39.489 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [Check for presence of ansible managed header, fingerprint] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_header.yml:9
Sunday 13 October 2024  03:55:22 -0400 (0:00:00.049)       0:00:39.538 ******** 
ok: [managed-node2] => {
    "changed": false
}

MSG:

All assertions passed

TASK [Cleanup] *****************************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:204
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.106)       0:00:39.645 ******** 
included: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml for managed-node2

TASK [Set platform/version specific variables] *********************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:2
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.069)       0:00:39.715 ******** 

TASK [fedora.linux_system_roles.vpn : Ensure ansible_facts used by role] *******
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:2
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.057)       0:00:39.772 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Check if system is ostree] ***************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:10
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.057)       0:00:39.830 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Set flag to indicate system is ostree] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:15
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.101)       0:00:39.932 ******** 
skipping: [managed-node2] => {
    "changed": false, 
    "skip_reason": "Conditional result was False"
}

TASK [fedora.linux_system_roles.vpn : Set platform/version specific variables] ***
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/set_vars.yml:19
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.050)       0:00:39.982 ******** 
skipping: [managed-node2] => (item=RedHat.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "RedHat.yml", 
    "skip_reason": "Conditional result was False"
}
skipping: [managed-node2] => (item=CentOS.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS.yml", 
    "skip_reason": "Conditional result was False"
}
ok: [managed-node2] => (item=CentOS_7.yml) => {
    "ansible_facts": {
        "__vpn_nss_location": "/etc/ipsec.d"
    }, 
    "ansible_included_var_files": [
        "/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/vars/CentOS_7.yml"
    ], 
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.yml"
}
skipping: [managed-node2] => (item=CentOS_7.9.yml)  => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "CentOS_7.9.yml", 
    "skip_reason": "Conditional result was False"
}

TASK [Get service facts] *******************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:8
Sunday 13 October 2024  03:55:23 -0400 (0:00:00.089)       0:00:40.072 ******** 
ok: [managed-node2] => {
    "ansible_facts": {
        "services": {
            "NetworkManager-dispatcher.service": {
                "name": "NetworkManager-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "NetworkManager-wait-online.service": {
                "name": "NetworkManager-wait-online.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "NetworkManager.service": {
                "name": "NetworkManager.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "arp-ethers.service": {
                "name": "arp-ethers.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "auditd.service": {
                "name": "auditd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "auth-rpcgss-module.service": {
                "name": "auth-rpcgss-module.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "autovt@.service": {
                "name": "autovt@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "blk-availability.service": {
                "name": "blk-availability.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "brandbot.service": {
                "name": "brandbot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "chrony-dnssrv@.service": {
                "name": "chrony-dnssrv@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "chrony-wait.service": {
                "name": "chrony-wait.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "chronyd.service": {
                "name": "chronyd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "cloud-config.service": {
                "name": "cloud-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-final.service": {
                "name": "cloud-final.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init-local.service": {
                "name": "cloud-init-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "cloud-init.service": {
                "name": "cloud-init.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "console-getty.service": {
                "name": "console-getty.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "console-shell.service": {
                "name": "console-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "container-getty@.service": {
                "name": "container-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "cpupower.service": {
                "name": "cpupower.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "crond.service": {
                "name": "crond.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "dbus-org.fedoraproject.FirewallD1.service": {
                "name": "dbus-org.fedoraproject.FirewallD1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.hostname1.service": {
                "name": "dbus-org.freedesktop.hostname1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.import1.service": {
                "name": "dbus-org.freedesktop.import1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.locale1.service": {
                "name": "dbus-org.freedesktop.locale1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.login1.service": {
                "name": "dbus-org.freedesktop.login1.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.machine1.service": {
                "name": "dbus-org.freedesktop.machine1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus-org.freedesktop.nm-dispatcher.service": {
                "name": "dbus-org.freedesktop.nm-dispatcher.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "dbus-org.freedesktop.timedate1.service": {
                "name": "dbus-org.freedesktop.timedate1.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "dbus.service": {
                "name": "dbus.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "debug-shell.service": {
                "name": "debug-shell.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "dracut-cmdline.service": {
                "name": "dracut-cmdline.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-initqueue.service": {
                "name": "dracut-initqueue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-mount.service": {
                "name": "dracut-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-mount.service": {
                "name": "dracut-pre-mount.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-pivot.service": {
                "name": "dracut-pre-pivot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-trigger.service": {
                "name": "dracut-pre-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-pre-udev.service": {
                "name": "dracut-pre-udev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "dracut-shutdown.service": {
                "name": "dracut-shutdown.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "ebtables.service": {
                "name": "ebtables.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "emergency.service": {
                "name": "emergency.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "firewalld.service": {
                "name": "firewalld.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "fstrim.service": {
                "name": "fstrim.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "getty@.service": {
                "name": "getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "enabled"
            }, 
            "getty@tty1.service": {
                "name": "getty@tty1.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "gssproxy.service": {
                "name": "gssproxy.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "disabled"
            }, 
            "halt-local.service": {
                "name": "halt-local.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "initrd-cleanup.service": {
                "name": "initrd-cleanup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-parse-etc.service": {
                "name": "initrd-parse-etc.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-switch-root.service": {
                "name": "initrd-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "initrd-udevadm-cleanup-db.service": {
                "name": "initrd-udevadm-cleanup-db.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "iprdump.service": {
                "name": "iprdump.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprinit.service": {
                "name": "iprinit.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "iprupdate.service": {
                "name": "iprupdate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "ipsec.service": {
                "name": "ipsec.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "irqbalance.service": {
                "name": "irqbalance.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "kdump.service": {
                "name": "kdump.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "kmod-static-nodes.service": {
                "name": "kmod-static-nodes.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "messagebus.service": {
                "name": "messagebus.service", 
                "source": "systemd", 
                "state": "active", 
                "status": "static"
            }, 
            "microcode.service": {
                "name": "microcode.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "netconsole": {
                "name": "netconsole", 
                "source": "sysv", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "network": {
                "name": "network", 
                "source": "sysv", 
                "state": "running", 
                "status": "enabled"
            }, 
            "network.service": {
                "name": "network.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "nfs-blkmap.service": {
                "name": "nfs-blkmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-config.service": {
                "name": "nfs-config.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-idmap.service": {
                "name": "nfs-idmap.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-idmapd.service": {
                "name": "nfs-idmapd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-lock.service": {
                "name": "nfs-lock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-mountd.service": {
                "name": "nfs-mountd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs-rquotad.service": {
                "name": "nfs-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfs-secure.service": {
                "name": "nfs-secure.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "nfs-server.service": {
                "name": "nfs-server.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "nfs-utils.service": {
                "name": "nfs-utils.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "nfs.service": {
                "name": "nfs.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "nfslock.service": {
                "name": "nfslock.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "plymouth-halt.service": {
                "name": "plymouth-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-kexec.service": {
                "name": "plymouth-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-poweroff.service": {
                "name": "plymouth-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-quit-wait.service": {
                "name": "plymouth-quit-wait.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-quit.service": {
                "name": "plymouth-quit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-read-write.service": {
                "name": "plymouth-read-write.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-reboot.service": {
                "name": "plymouth-reboot.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "plymouth-start.service": {
                "name": "plymouth-start.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "disabled"
            }, 
            "plymouth-switch-root.service": {
                "name": "plymouth-switch-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "polkit.service": {
                "name": "polkit.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "postfix.service": {
                "name": "postfix.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "qemu-guest-agent.service": {
                "name": "qemu-guest-agent.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "quotaon.service": {
                "name": "quotaon.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rc-local.service": {
                "name": "rc-local.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rdisc.service": {
                "name": "rdisc.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rescue.service": {
                "name": "rescue.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "restraintd.service": {
                "name": "restraintd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rhel-autorelabel-mark.service": {
                "name": "rhel-autorelabel-mark.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-autorelabel.service": {
                "name": "rhel-autorelabel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-configure.service": {
                "name": "rhel-configure.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-dmesg.service": {
                "name": "rhel-dmesg.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-domainname.service": {
                "name": "rhel-domainname.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-import-state.service": {
                "name": "rhel-import-state.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-loadmodules.service": {
                "name": "rhel-loadmodules.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rhel-readonly.service": {
                "name": "rhel-readonly.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "rngd.service": {
                "name": "rngd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpc-gssd.service": {
                "name": "rpc-gssd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-rquotad.service": {
                "name": "rpc-rquotad.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rpc-statd-notify.service": {
                "name": "rpc-statd-notify.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpc-statd.service": {
                "name": "rpc-statd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "rpcbind.service": {
                "name": "rpcbind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "rpcgssd.service": {
                "name": "rpcgssd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rpcidmapd.service": {
                "name": "rpcidmapd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "rsyncd.service": {
                "name": "rsyncd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "rsyncd@.service": {
                "name": "rsyncd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "rsyslog.service": {
                "name": "rsyslog.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "selinux-policy-migrate-local-changes@.service": {
                "name": "selinux-policy-migrate-local-changes@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "selinux-policy-migrate-local-changes@targeted.service": {
                "name": "selinux-policy-migrate-local-changes@targeted.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "unknown"
            }, 
            "serial-getty@.service": {
                "name": "serial-getty@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "serial-getty@ttyS0.service": {
                "name": "serial-getty@ttyS0.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "unknown"
            }, 
            "sshd-keygen.service": {
                "name": "sshd-keygen.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "sshd.service": {
                "name": "sshd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "sshd@.service": {
                "name": "sshd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-ask-password-console.service": {
                "name": "systemd-ask-password-console.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-plymouth.service": {
                "name": "systemd-ask-password-plymouth.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-ask-password-wall.service": {
                "name": "systemd-ask-password-wall.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-backlight@.service": {
                "name": "systemd-backlight@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-binfmt.service": {
                "name": "systemd-binfmt.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-bootchart.service": {
                "name": "systemd-bootchart.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }, 
            "systemd-firstboot.service": {
                "name": "systemd-firstboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck-root.service": {
                "name": "systemd-fsck-root.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-fsck@.service": {
                "name": "systemd-fsck@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-halt.service": {
                "name": "systemd-halt.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hibernate-resume@.service": {
                "name": "systemd-hibernate-resume@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-hibernate.service": {
                "name": "systemd-hibernate.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hostnamed.service": {
                "name": "systemd-hostnamed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-hwdb-update.service": {
                "name": "systemd-hwdb-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-hybrid-sleep.service": {
                "name": "systemd-hybrid-sleep.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-importd.service": {
                "name": "systemd-importd.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-initctl.service": {
                "name": "systemd-initctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-catalog-update.service": {
                "name": "systemd-journal-catalog-update.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journal-flush.service": {
                "name": "systemd-journal-flush.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-journald.service": {
                "name": "systemd-journald.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-kexec.service": {
                "name": "systemd-kexec.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-localed.service": {
                "name": "systemd-localed.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-logind.service": {
                "name": "systemd-logind.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-machine-id-commit.service": {
                "name": "systemd-machine-id-commit.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-machined.service": {
                "name": "systemd-machined.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-modules-load.service": {
                "name": "systemd-modules-load.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-nspawn@.service": {
                "name": "systemd-nspawn@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "disabled"
            }, 
            "systemd-poweroff.service": {
                "name": "systemd-poweroff.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-quotacheck.service": {
                "name": "systemd-quotacheck.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-random-seed.service": {
                "name": "systemd-random-seed.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-readahead-collect.service": {
                "name": "systemd-readahead-collect.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-readahead-done.service": {
                "name": "systemd-readahead-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "indirect"
            }, 
            "systemd-readahead-drop.service": {
                "name": "systemd-readahead-drop.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "enabled"
            }, 
            "systemd-readahead-replay.service": {
                "name": "systemd-readahead-replay.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "enabled"
            }, 
            "systemd-reboot.service": {
                "name": "systemd-reboot.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-remount-fs.service": {
                "name": "systemd-remount-fs.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-rfkill@.service": {
                "name": "systemd-rfkill@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "systemd-shutdownd.service": {
                "name": "systemd-shutdownd.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-suspend.service": {
                "name": "systemd-suspend.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-sysctl.service": {
                "name": "systemd-sysctl.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-timedated.service": {
                "name": "systemd-timedated.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-tmpfiles-clean.service": {
                "name": "systemd-tmpfiles-clean.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup-dev.service": {
                "name": "systemd-tmpfiles-setup-dev.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-tmpfiles-setup.service": {
                "name": "systemd-tmpfiles-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udev-settle.service": {
                "name": "systemd-udev-settle.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "static"
            }, 
            "systemd-udev-trigger.service": {
                "name": "systemd-udev-trigger.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-udevd.service": {
                "name": "systemd-udevd.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "static"
            }, 
            "systemd-update-done.service": {
                "name": "systemd-update-done.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp-runlevel.service": {
                "name": "systemd-update-utmp-runlevel.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-update-utmp.service": {
                "name": "systemd-update-utmp.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-user-sessions.service": {
                "name": "systemd-user-sessions.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "systemd-vconsole-setup.service": {
                "name": "systemd-vconsole-setup.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "teamd@.service": {
                "name": "teamd@.service", 
                "source": "systemd", 
                "state": "unknown", 
                "status": "static"
            }, 
            "tuned.service": {
                "name": "tuned.service", 
                "source": "systemd", 
                "state": "running", 
                "status": "enabled"
            }, 
            "unbound-anchor.service": {
                "name": "unbound-anchor.service", 
                "source": "systemd", 
                "state": "stopped", 
                "status": "static"
            }, 
            "wpa_supplicant.service": {
                "name": "wpa_supplicant.service", 
                "source": "systemd", 
                "state": "inactive", 
                "status": "disabled"
            }
        }
    }, 
    "changed": false
}

TASK [Stop ipsec service] ******************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:11
Sunday 13 October 2024  03:55:24 -0400 (0:00:00.916)       0:00:40.989 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "enabled": false, 
    "name": "ipsec", 
    "state": "stopped", 
    "status": {
        "ActiveEnterTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "ActiveEnterTimestampMonotonic": "788026572", 
        "ActiveExitTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "ActiveExitTimestampMonotonic": "787551931", 
        "ActiveState": "active", 
        "After": "systemd-journald.socket system.slice basic.target network-online.target", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "yes", 
        "AssertTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "AssertTimestampMonotonic": "787607081", 
        "Before": "shutdown.target multi-user.target", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "yes", 
        "ConditionTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "ConditionTimestampMonotonic": "787607080", 
        "Conflicts": "shutdown.target", 
        "ControlGroup": "/system.slice/ipsec.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "Internet Key Exchange (IKE) Protocol Daemon for IPsec", 
        "DevicePolicy": "auto", 
        "Documentation": "man:ipsec(8) man:pluto(8) man:ipsec.conf(5)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "11811", 
        "ExecMainStartTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "ExecMainStartTimestampMonotonic": "787890730", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/usr/libexec/ipsec/whack ; argv[]=/usr/libexec/ipsec/whack --listen ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/libexec/ipsec/pluto ; argv[]=/usr/libexec/ipsec/pluto --leak-detective --config /etc/ipsec.conf --nofork ; ignore_errors=no ; start_time=[Sun 2024-10-13 03:55:14 EDT] ; stop_time=[n/a] ; pid=11811 ; code=(null) ; status=0/0 }", 
        "ExecStartPre": "{ path=/usr/sbin/ipsec ; argv[]=/usr/sbin/ipsec --checknflog ; ignore_errors=no ; start_time=[Sun 2024-10-13 03:55:14 EDT] ; stop_time=[Sun 2024-10-13 03:55:14 EDT] ; pid=11800 ; code=exited ; status=0 }", 
        "ExecStop": "{ path=/usr/libexec/ipsec/whack ; argv[]=/usr/libexec/ipsec/whack --shutdown ; ignore_errors=no ; start_time=[Sun 2024-10-13 03:55:14 EDT] ; stop_time=[Sun 2024-10-13 03:55:14 EDT] ; pid=11505 ; code=exited ; status=0 }", 
        "ExecStopPost": "{ path=/usr/sbin/ipsec ; argv[]=/usr/sbin/ipsec --stopnflog ; ignore_errors=no ; start_time=[Sun 2024-10-13 03:55:14 EDT] ; stop_time=[Sun 2024-10-13 03:55:14 EDT] ; pid=11523 ; code=exited ; status=0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/ipsec.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "ipsec.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "InactiveEnterTimestampMonotonic": "787606271", 
        "InactiveExitTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "InactiveExitTimestampMonotonic": "787607607", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "control-group", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "11811", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "ipsec.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "all", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Requires": "basic.target system.slice", 
        "Restart": "always", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "inherit", 
        "StandardInput": "null", 
        "StandardOutput": "journal", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StatusText": "Startup completed.", 
        "StopWhenUnneeded": "no", 
        "SubState": "running", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "notify", 
        "UMask": "0022", 
        "UnitFilePreset": "disabled", 
        "UnitFileState": "enabled", 
        "WantedBy": "multi-user.target", 
        "Wants": "network-online.target", 
        "WatchdogTimestamp": "Sun 2024-10-13 03:55:14 EDT", 
        "WatchdogTimestampMonotonic": "788026527", 
        "WatchdogUSec": "3min 20s"
    }
}

TASK [Uninstall packages] ******************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:18
Sunday 13 October 2024  03:55:24 -0400 (0:00:00.577)       0:00:41.567 ******** 
changed: [managed-node2] => (item=libreswan) => {
    "ansible_loop_var": "item", 
    "changed": true, 
    "changes": {
        "removed": [
            "libreswan"
        ]
    }, 
    "item": "libreswan", 
    "rc": 0, 
    "results": [
        "Loaded plugins: fastestmirror\nResolving Dependencies\n--> Running transaction check\n---> Package libreswan.x86_64 0:3.25-9.1.el7_8 will be erased\n--> Finished Dependency Resolution\n\nDependencies Resolved\n\n================================================================================\n Package           Arch           Version                Repository        Size\n================================================================================\nRemoving:\n libreswan         x86_64         3.25-9.1.el7_8         @updates         4.7 M\n\nTransaction Summary\n================================================================================\nRemove  1 Package\n\nInstalled size: 4.7 M\nDownloading packages:\nRunning transaction check\nRunning transaction test\nTransaction test succeeded\nRunning transaction\n  Erasing    : libreswan-3.25-9.1.el7_8.x86_64                              1/1 \nwarning: /etc/ipsec.d/policies/private-or-clear saved as /etc/ipsec.d/policies/private-or-clear.rpmsave\nwarning: /etc/ipsec.d/policies/private saved as /etc/ipsec.d/policies/private.rpmsave\nwarning: /etc/ipsec.d/policies/clear saved as /etc/ipsec.d/policies/clear.rpmsave\n  Verifying  : libreswan-3.25-9.1.el7_8.x86_64                              1/1 \n\nRemoved:\n  libreswan.x86_64 0:3.25-9.1.el7_8                                             \n\nComplete!\n"
    ]
}

TASK [Remove ipsec.conf files] *************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:43
Sunday 13 October 2024  03:55:26 -0400 (0:00:01.584)       0:00:43.152 ******** 
ok: [managed-node2] => (item=host01.local) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "host01.local", 
    "path": "/etc/ipsec.d/managed-node2-to-host01.local.conf", 
    "state": "absent"
}
ok: [managed-node2] => (item=host02.local) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "host02.local", 
    "path": "/etc/ipsec.d/managed-node2-to-host02.local.conf", 
    "state": "absent"
}
ok: [managed-node2] => (item=managed-node1) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "managed-node1", 
    "path": "/etc/ipsec.d/managed-node2-to-managed-node1.conf", 
    "state": "absent"
}
ok: [managed-node2] => (item=managed-node3) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "managed-node3", 
    "path": "/etc/ipsec.d/managed-node2-to-managed-node3.conf", 
    "state": "absent"
}

TASK [Remove ipsec.secrets files] **********************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:49
Sunday 13 October 2024  03:55:27 -0400 (0:00:01.232)       0:00:44.384 ******** 
ok: [managed-node2] => (item=host01.local) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "host01.local", 
    "path": "/etc/ipsec.d/managed-node2-to-host01.local.secrets", 
    "state": "absent"
}
ok: [managed-node2] => (item=host02.local) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "host02.local", 
    "path": "/etc/ipsec.d/managed-node2-to-host02.local.secrets", 
    "state": "absent"
}
ok: [managed-node2] => (item=managed-node1) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "managed-node1", 
    "path": "/etc/ipsec.d/managed-node2-to-managed-node1.secrets", 
    "state": "absent"
}
ok: [managed-node2] => (item=managed-node3) => {
    "ansible_loop_var": "item", 
    "changed": false, 
    "item": "managed-node3", 
    "path": "/etc/ipsec.d/managed-node2-to-managed-node3.secrets", 
    "state": "absent"
}

TASK [Stop firewall] ***********************************************************
task path: /tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:55
Sunday 13 October 2024  03:55:29 -0400 (0:00:01.275)       0:00:45.660 ******** 
changed: [managed-node2] => {
    "changed": true, 
    "name": "firewalld", 
    "state": "stopped", 
    "status": {
        "ActiveEnterTimestamp": "Sun 2024-10-13 03:55:00 EDT", 
        "ActiveEnterTimestampMonotonic": "773316465", 
        "ActiveExitTimestampMonotonic": "0", 
        "ActiveState": "active", 
        "After": "basic.target dbus.service system.slice polkit.service", 
        "AllowIsolate": "no", 
        "AmbientCapabilities": "0", 
        "AssertResult": "yes", 
        "AssertTimestamp": "Sun 2024-10-13 03:54:59 EDT", 
        "AssertTimestampMonotonic": "773073776", 
        "Before": "shutdown.target multi-user.target network-pre.target", 
        "BlockIOAccounting": "no", 
        "BlockIOWeight": "18446744073709551615", 
        "BusName": "org.fedoraproject.FirewallD1", 
        "CPUAccounting": "no", 
        "CPUQuotaPerSecUSec": "infinity", 
        "CPUSchedulingPolicy": "0", 
        "CPUSchedulingPriority": "0", 
        "CPUSchedulingResetOnFork": "no", 
        "CPUShares": "18446744073709551615", 
        "CanIsolate": "no", 
        "CanReload": "yes", 
        "CanStart": "yes", 
        "CanStop": "yes", 
        "CapabilityBoundingSet": "18446744073709551615", 
        "CollectMode": "inactive", 
        "ConditionResult": "yes", 
        "ConditionTimestamp": "Sun 2024-10-13 03:54:59 EDT", 
        "ConditionTimestampMonotonic": "773073774", 
        "Conflicts": "ip6tables.service shutdown.target iptables.service ebtables.service ipset.service", 
        "ControlGroup": "/system.slice/firewalld.service", 
        "ControlPID": "0", 
        "DefaultDependencies": "yes", 
        "Delegate": "no", 
        "Description": "firewalld - dynamic firewall daemon", 
        "DevicePolicy": "auto", 
        "Documentation": "man:firewalld(1)", 
        "EnvironmentFile": "/etc/sysconfig/firewalld (ignore_errors=yes)", 
        "ExecMainCode": "0", 
        "ExecMainExitTimestampMonotonic": "0", 
        "ExecMainPID": "10021", 
        "ExecMainStartTimestamp": "Sun 2024-10-13 03:54:59 EDT", 
        "ExecMainStartTimestampMonotonic": "773075043", 
        "ExecMainStatus": "0", 
        "ExecReload": "{ path=/bin/kill ; argv[]=/bin/kill -HUP $MAINPID ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "ExecStart": "{ path=/usr/sbin/firewalld ; argv[]=/usr/sbin/firewalld --nofork --nopid $FIREWALLD_ARGS ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", 
        "FailureAction": "none", 
        "FileDescriptorStoreMax": "0", 
        "FragmentPath": "/usr/lib/systemd/system/firewalld.service", 
        "GuessMainPID": "yes", 
        "IOScheduling": "0", 
        "Id": "firewalld.service", 
        "IgnoreOnIsolate": "no", 
        "IgnoreOnSnapshot": "no", 
        "IgnoreSIGPIPE": "yes", 
        "InactiveEnterTimestampMonotonic": "0", 
        "InactiveExitTimestamp": "Sun 2024-10-13 03:54:59 EDT", 
        "InactiveExitTimestampMonotonic": "773075080", 
        "JobTimeoutAction": "none", 
        "JobTimeoutUSec": "0", 
        "KillMode": "mixed", 
        "KillSignal": "15", 
        "LimitAS": "18446744073709551615", 
        "LimitCORE": "18446744073709551615", 
        "LimitCPU": "18446744073709551615", 
        "LimitDATA": "18446744073709551615", 
        "LimitFSIZE": "18446744073709551615", 
        "LimitLOCKS": "18446744073709551615", 
        "LimitMEMLOCK": "65536", 
        "LimitMSGQUEUE": "819200", 
        "LimitNICE": "0", 
        "LimitNOFILE": "4096", 
        "LimitNPROC": "14311", 
        "LimitRSS": "18446744073709551615", 
        "LimitRTPRIO": "0", 
        "LimitRTTIME": "18446744073709551615", 
        "LimitSIGPENDING": "14311", 
        "LimitSTACK": "18446744073709551615", 
        "LoadState": "loaded", 
        "MainPID": "10021", 
        "MemoryAccounting": "no", 
        "MemoryCurrent": "18446744073709551615", 
        "MemoryLimit": "18446744073709551615", 
        "MountFlags": "0", 
        "Names": "firewalld.service", 
        "NeedDaemonReload": "no", 
        "Nice": "0", 
        "NoNewPrivileges": "no", 
        "NonBlocking": "no", 
        "NotifyAccess": "none", 
        "OOMScoreAdjust": "0", 
        "OnFailureJobMode": "replace", 
        "PermissionsStartOnly": "no", 
        "PrivateDevices": "no", 
        "PrivateNetwork": "no", 
        "PrivateTmp": "no", 
        "ProtectHome": "no", 
        "ProtectSystem": "no", 
        "RefuseManualStart": "no", 
        "RefuseManualStop": "no", 
        "RemainAfterExit": "no", 
        "Requires": "basic.target system.slice", 
        "Restart": "no", 
        "RestartUSec": "100ms", 
        "Result": "success", 
        "RootDirectoryStartOnly": "no", 
        "RuntimeDirectoryMode": "0755", 
        "SameProcessGroup": "no", 
        "SecureBits": "0", 
        "SendSIGHUP": "no", 
        "SendSIGKILL": "yes", 
        "Slice": "system.slice", 
        "StandardError": "null", 
        "StandardInput": "null", 
        "StandardOutput": "null", 
        "StartLimitAction": "none", 
        "StartLimitBurst": "5", 
        "StartLimitInterval": "10000000", 
        "StartupBlockIOWeight": "18446744073709551615", 
        "StartupCPUShares": "18446744073709551615", 
        "StatusErrno": "0", 
        "StopWhenUnneeded": "no", 
        "SubState": "running", 
        "SyslogLevelPrefix": "yes", 
        "SyslogPriority": "30", 
        "SystemCallErrorNumber": "0", 
        "TTYReset": "no", 
        "TTYVHangup": "no", 
        "TTYVTDisallocate": "no", 
        "TasksAccounting": "no", 
        "TasksCurrent": "18446744073709551615", 
        "TasksMax": "18446744073709551615", 
        "TimeoutStartUSec": "1min 30s", 
        "TimeoutStopUSec": "1min 30s", 
        "TimerSlackNSec": "50000", 
        "Transient": "no", 
        "Type": "dbus", 
        "UMask": "0022", 
        "UnitFilePreset": "enabled", 
        "UnitFileState": "enabled", 
        "WantedBy": "multi-user.target", 
        "Wants": "network-pre.target", 
        "WatchdogTimestamp": "Sun 2024-10-13 03:55:00 EDT", 
        "WatchdogTimestampMonotonic": "773316372", 
        "WatchdogUSec": "0"
    }
}
META: ran handlers
META: ran handlers

PLAY RECAP *********************************************************************
managed-node2              : ok=87   changed=16   unreachable=0    failed=0    skipped=48   rescued=0    ignored=0   

Sunday 13 October 2024  03:55:30 -0400 (0:00:01.198)       0:00:46.859 ******** 
=============================================================================== 
fedora.linux_system_roles.vpn : Ensure required packages are installed --- 7.00s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/main.yml:6 
fedora.linux_system_roles.vpn : Wait for ssh connection to return ------- 3.91s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:11 
fedora.linux_system_roles.vpn : Wait for ssh connection to return ------- 3.60s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:11 
fedora.linux_system_roles.vpn : Write tunnel policies for each network --- 2.25s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:48 
Uninstall packages ------------------------------------------------------ 1.58s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:18 
Remove ipsec.secrets files ---------------------------------------------- 1.28s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:49 
Gathering Facts --------------------------------------------------------- 1.24s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tests_mesh_cert.yml:3 
Remove ipsec.conf files ------------------------------------------------- 1.23s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:43 
Stop firewall ----------------------------------------------------------- 1.20s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:55 
fedora.linux_system_roles.vpn : Populate service facts ------------------ 1.14s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/selinux.yml:3 
fedora.linux_system_roles.firewall : Configure firewall ----------------- 0.96s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:71 
Get service facts ------------------------------------------------------- 0.92s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/cleanup.yml:8 
fedora.linux_system_roles.vpn : Enable and restart vpn services --------- 0.87s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:2 
fedora.linux_system_roles.firewall : Enable and start firewalld service --- 0.86s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:28 
fedora.linux_system_roles.firewall : Unmask firewalld service ----------- 0.80s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/main.yml:22 
fedora.linux_system_roles.vpn : Deploy opportunistic configuration to each node --- 0.80s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/mesh_conf.yml:58 
Get firewall service ---------------------------------------------------- 0.67s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml:9 
fedora.linux_system_roles.vpn : Enable and restart vpn services --------- 0.62s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/vpn/tasks/enable_restart_vpn.yml:2 
fedora.linux_system_roles.firewall : Install firewalld ------------------ 0.59s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/roles/firewall/tasks/firewalld.yml:31 
Get associated selinux ports -------------------------------------------- 0.58s
/tmp/collections-vqX/ansible_collections/fedora/linux_system_roles/tests/vpn/tasks/check_firewall_selinux.yml:19