mirror of
https://github.com/ansible-collections/community.general.git
synced 2025-04-18 00:11:26 -07:00
[stable-7] Revert new features to be able to do 7.2.1 release (#7042)
* Revert "[PR #7020/b46d5d81 backport][stable-7] redfish_utils: Add support for "nextLink" property tag pagination (#7026)" This reverts commit1dad95370e
. * Revert "[PR #6914/17b4219b backport][stable-7] proxmox_kvm: enable 'force' restart of vm (as documented) (#6997)" This reverts commit7d68af57af
. * Revert "[PR #6976/d7c1a814 backport][stable-7] [proxmox_vm_info] Re-use cluster resources API to use module without requiring node param (#6993)" This reverts commitfb3768aada
.
This commit is contained in:
parent
eda3d160fa
commit
7cf834fb3c
7 changed files with 146 additions and 423 deletions
|
@ -1,2 +0,0 @@
|
||||||
minor_changes:
|
|
||||||
- proxmox_kvm - enabled force restart of VM, bringing the ``force`` parameter functionality in line with what is described in the docs (https://github.com/ansible-collections/community.general/pull/6914).
|
|
|
@ -1,2 +0,0 @@
|
||||||
minor_changes:
|
|
||||||
- proxmox_vm_info - ``node`` parameter is no longer required. Information can be obtained for the whole cluster (https://github.com/ansible-collections/community.general/pull/6976).
|
|
|
@ -1,2 +0,0 @@
|
||||||
minor_changes:
|
|
||||||
- redfish_utils module utils - add support for following ``@odata.nextLink`` pagination in ``software_inventory`` responses feature (https://github.com/ansible-collections/community.general/pull/7020).
|
|
|
@ -1499,37 +1499,29 @@ class RedfishUtils(object):
|
||||||
|
|
||||||
def _software_inventory(self, uri):
|
def _software_inventory(self, uri):
|
||||||
result = {}
|
result = {}
|
||||||
result['entries'] = []
|
response = self.get_request(self.root_uri + uri)
|
||||||
|
if response['ret'] is False:
|
||||||
|
return response
|
||||||
|
result['ret'] = True
|
||||||
|
data = response['data']
|
||||||
|
|
||||||
while uri:
|
result['entries'] = []
|
||||||
response = self.get_request(self.root_uri + uri)
|
for member in data[u'Members']:
|
||||||
|
uri = self.root_uri + member[u'@odata.id']
|
||||||
|
# Get details for each software or firmware member
|
||||||
|
response = self.get_request(uri)
|
||||||
if response['ret'] is False:
|
if response['ret'] is False:
|
||||||
return response
|
return response
|
||||||
result['ret'] = True
|
result['ret'] = True
|
||||||
|
|
||||||
data = response['data']
|
data = response['data']
|
||||||
if data.get('Members@odata.nextLink'):
|
software = {}
|
||||||
uri = data.get('Members@odata.nextLink')
|
# Get these standard properties if present
|
||||||
else:
|
for key in ['Name', 'Id', 'Status', 'Version', 'Updateable',
|
||||||
uri = None
|
'SoftwareId', 'LowestSupportedVersion', 'Manufacturer',
|
||||||
|
'ReleaseDate']:
|
||||||
for member in data[u'Members']:
|
if key in data:
|
||||||
fw_uri = self.root_uri + member[u'@odata.id']
|
software[key] = data.get(key)
|
||||||
# Get details for each software or firmware member
|
result['entries'].append(software)
|
||||||
response = self.get_request(fw_uri)
|
|
||||||
if response['ret'] is False:
|
|
||||||
return response
|
|
||||||
result['ret'] = True
|
|
||||||
data = response['data']
|
|
||||||
software = {}
|
|
||||||
# Get these standard properties if present
|
|
||||||
for key in ['Name', 'Id', 'Status', 'Version', 'Updateable',
|
|
||||||
'SoftwareId', 'LowestSupportedVersion', 'Manufacturer',
|
|
||||||
'ReleaseDate']:
|
|
||||||
if key in data:
|
|
||||||
software[key] = data.get(key)
|
|
||||||
result['entries'].append(software)
|
|
||||||
|
|
||||||
return result
|
return result
|
||||||
|
|
||||||
def get_firmware_inventory(self):
|
def get_firmware_inventory(self):
|
||||||
|
|
|
@ -1111,11 +1111,11 @@ class ProxmoxKvmAnsible(ProxmoxAnsible):
|
||||||
return False
|
return False
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def restart_vm(self, vm, force, **status):
|
def restart_vm(self, vm, **status):
|
||||||
vmid = vm['vmid']
|
vmid = vm['vmid']
|
||||||
try:
|
try:
|
||||||
proxmox_node = self.proxmox_api.nodes(vm['node'])
|
proxmox_node = self.proxmox_api.nodes(vm['node'])
|
||||||
taskid = proxmox_node.qemu(vmid).status.reset.post() if force else proxmox_node.qemu(vmid).status.reboot.post()
|
taskid = proxmox_node.qemu(vmid).status.reboot.post()
|
||||||
if not self.wait_for_task(vm['node'], taskid):
|
if not self.wait_for_task(vm['node'], taskid):
|
||||||
self.module.fail_json(msg='Reached timeout while waiting for rebooting VM. Last line in task before timeout: %s' %
|
self.module.fail_json(msg='Reached timeout while waiting for rebooting VM. Last line in task before timeout: %s' %
|
||||||
proxmox_node.tasks(taskid).log.get()[:1])
|
proxmox_node.tasks(taskid).log.get()[:1])
|
||||||
|
@ -1493,7 +1493,7 @@ def main():
|
||||||
if vm['status'] == 'stopped':
|
if vm['status'] == 'stopped':
|
||||||
module.exit_json(changed=False, vmid=vmid, msg="VM %s is not running" % vmid, **status)
|
module.exit_json(changed=False, vmid=vmid, msg="VM %s is not running" % vmid, **status)
|
||||||
|
|
||||||
if proxmox.restart_vm(vm, force=module.params['force']):
|
if proxmox.restart_vm(vm):
|
||||||
module.exit_json(changed=True, vmid=vmid, msg="VM %s is restarted" % vmid, **status)
|
module.exit_json(changed=True, vmid=vmid, msg="VM %s is restarted" % vmid, **status)
|
||||||
|
|
||||||
elif state == 'absent':
|
elif state == 'absent':
|
||||||
|
|
|
@ -20,7 +20,8 @@ author: 'Sergei Antipov (@UnderGreen) <greendayonfire at gmail dot com>'
|
||||||
options:
|
options:
|
||||||
node:
|
node:
|
||||||
description:
|
description:
|
||||||
- Restrict results to a specific Proxmox VE node.
|
- Node where to get virtual machines info.
|
||||||
|
required: true
|
||||||
type: str
|
type: str
|
||||||
type:
|
type:
|
||||||
description:
|
description:
|
||||||
|
@ -96,18 +97,14 @@ proxmox_vms:
|
||||||
"disk": 0,
|
"disk": 0,
|
||||||
"diskread": 0,
|
"diskread": 0,
|
||||||
"diskwrite": 0,
|
"diskwrite": 0,
|
||||||
"id": "qemu/100",
|
|
||||||
"maxcpu": 1,
|
|
||||||
"maxdisk": 34359738368,
|
"maxdisk": 34359738368,
|
||||||
"maxmem": 4294967296,
|
"maxmem": 4294967296,
|
||||||
"mem": 35158379,
|
"mem": 35158379,
|
||||||
"name": "pxe.home.arpa",
|
"name": "pxe.home.arpa",
|
||||||
"netin": 99715803,
|
"netin": 99715803,
|
||||||
"netout": 14237835,
|
"netout": 14237835,
|
||||||
"node": "pve",
|
|
||||||
"pid": 1947197,
|
"pid": 1947197,
|
||||||
"status": "running",
|
"status": "running",
|
||||||
"template": False,
|
|
||||||
"type": "qemu",
|
"type": "qemu",
|
||||||
"uptime": 135530,
|
"uptime": 135530,
|
||||||
"vmid": 100
|
"vmid": 100
|
||||||
|
@ -118,17 +115,13 @@ proxmox_vms:
|
||||||
"disk": 0,
|
"disk": 0,
|
||||||
"diskread": 0,
|
"diskread": 0,
|
||||||
"diskwrite": 0,
|
"diskwrite": 0,
|
||||||
"id": "qemu/101",
|
|
||||||
"maxcpu": 1,
|
|
||||||
"maxdisk": 0,
|
"maxdisk": 0,
|
||||||
"maxmem": 536870912,
|
"maxmem": 536870912,
|
||||||
"mem": 0,
|
"mem": 0,
|
||||||
"name": "test1",
|
"name": "test1",
|
||||||
"netin": 0,
|
"netin": 0,
|
||||||
"netout": 0,
|
"netout": 0,
|
||||||
"node": "pve",
|
|
||||||
"status": "stopped",
|
"status": "stopped",
|
||||||
"template": False,
|
|
||||||
"type": "qemu",
|
"type": "qemu",
|
||||||
"uptime": 0,
|
"uptime": 0,
|
||||||
"vmid": 101
|
"vmid": 101
|
||||||
|
@ -140,54 +133,30 @@ from ansible.module_utils.basic import AnsibleModule
|
||||||
from ansible_collections.community.general.plugins.module_utils.proxmox import (
|
from ansible_collections.community.general.plugins.module_utils.proxmox import (
|
||||||
proxmox_auth_argument_spec,
|
proxmox_auth_argument_spec,
|
||||||
ProxmoxAnsible,
|
ProxmoxAnsible,
|
||||||
proxmox_to_ansible_bool,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
class ProxmoxVmInfoAnsible(ProxmoxAnsible):
|
class ProxmoxVmInfoAnsible(ProxmoxAnsible):
|
||||||
def get_vms_from_cluster_resources(self):
|
def get_qemu_vms(self, node, vmid=None):
|
||||||
try:
|
try:
|
||||||
return self.proxmox_api.cluster().resources().get(type="vm")
|
vms = self.proxmox_api.nodes(node).qemu().get()
|
||||||
except Exception as e:
|
for vm in vms:
|
||||||
self.module.fail_json(
|
vm["vmid"] = int(vm["vmid"])
|
||||||
msg="Failed to retrieve VMs information from cluster resources: %s" % e
|
vm["type"] = "qemu"
|
||||||
)
|
if vmid is None:
|
||||||
|
return vms
|
||||||
def get_vms_from_nodes(self, vms_unfiltered, type, vmid=None, node=None):
|
return [vm for vm in vms if vm["vmid"] == vmid]
|
||||||
vms = []
|
|
||||||
for vm in vms_unfiltered:
|
|
||||||
if (
|
|
||||||
type != vm["type"]
|
|
||||||
or (node and vm["node"] != node)
|
|
||||||
or (vmid and int(vm["vmid"]) != vmid)
|
|
||||||
):
|
|
||||||
continue
|
|
||||||
vms.append(vm)
|
|
||||||
nodes = frozenset([vm["node"] for vm in vms])
|
|
||||||
for node in nodes:
|
|
||||||
if type == "qemu":
|
|
||||||
vms_from_nodes = self.proxmox_api.nodes(node).qemu().get()
|
|
||||||
else:
|
|
||||||
vms_from_nodes = self.proxmox_api.nodes(node).lxc().get()
|
|
||||||
for vmn in vms_from_nodes:
|
|
||||||
for vm in vms:
|
|
||||||
if int(vm["vmid"]) == int(vmn["vmid"]):
|
|
||||||
vm.update(vmn)
|
|
||||||
vm["vmid"] = int(vm["vmid"])
|
|
||||||
vm["template"] = proxmox_to_ansible_bool(vm["template"])
|
|
||||||
break
|
|
||||||
|
|
||||||
return vms
|
|
||||||
|
|
||||||
def get_qemu_vms(self, vms_unfiltered, vmid=None, node=None):
|
|
||||||
try:
|
|
||||||
return self.get_vms_from_nodes(vms_unfiltered, "qemu", vmid, node)
|
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
self.module.fail_json(msg="Failed to retrieve QEMU VMs information: %s" % e)
|
self.module.fail_json(msg="Failed to retrieve QEMU VMs information: %s" % e)
|
||||||
|
|
||||||
def get_lxc_vms(self, vms_unfiltered, vmid=None, node=None):
|
def get_lxc_vms(self, node, vmid=None):
|
||||||
try:
|
try:
|
||||||
return self.get_vms_from_nodes(vms_unfiltered, "lxc", vmid, node)
|
vms = self.proxmox_api.nodes(node).lxc().get()
|
||||||
|
for vm in vms:
|
||||||
|
vm["vmid"] = int(vm["vmid"])
|
||||||
|
if vmid is None:
|
||||||
|
return vms
|
||||||
|
return [vm for vm in vms if vm["vmid"] == vmid]
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
self.module.fail_json(msg="Failed to retrieve LXC VMs information: %s" % e)
|
self.module.fail_json(msg="Failed to retrieve LXC VMs information: %s" % e)
|
||||||
|
|
||||||
|
@ -195,7 +164,7 @@ class ProxmoxVmInfoAnsible(ProxmoxAnsible):
|
||||||
def main():
|
def main():
|
||||||
module_args = proxmox_auth_argument_spec()
|
module_args = proxmox_auth_argument_spec()
|
||||||
vm_info_args = dict(
|
vm_info_args = dict(
|
||||||
node=dict(type="str", required=False),
|
node=dict(type="str", required=True),
|
||||||
type=dict(
|
type=dict(
|
||||||
type="str", choices=["lxc", "qemu", "all"], default="all", required=False
|
type="str", choices=["lxc", "qemu", "all"], default="all", required=False
|
||||||
),
|
),
|
||||||
|
@ -219,32 +188,27 @@ def main():
|
||||||
|
|
||||||
result = dict(changed=False)
|
result = dict(changed=False)
|
||||||
|
|
||||||
if node and proxmox.get_node(node) is None:
|
if proxmox.get_node(node) is None:
|
||||||
module.fail_json(msg="Node %s doesn't exist in PVE cluster" % node)
|
module.fail_json(msg="Node %s doesn't exist in PVE cluster" % node)
|
||||||
|
|
||||||
if not vmid and name:
|
if not vmid and name:
|
||||||
vmid = int(proxmox.get_vmid(name, ignore_missing=False))
|
vmid = int(proxmox.get_vmid(name, ignore_missing=False))
|
||||||
|
|
||||||
vms_cluster_resources = proxmox.get_vms_from_cluster_resources()
|
|
||||||
vms = None
|
vms = None
|
||||||
|
|
||||||
if type == "lxc":
|
if type == "lxc":
|
||||||
vms = proxmox.get_lxc_vms(vms_cluster_resources, vmid, node)
|
vms = proxmox.get_lxc_vms(node, vmid=vmid)
|
||||||
elif type == "qemu":
|
elif type == "qemu":
|
||||||
vms = proxmox.get_qemu_vms(vms_cluster_resources, vmid, node)
|
vms = proxmox.get_qemu_vms(node, vmid=vmid)
|
||||||
else:
|
else:
|
||||||
vms = proxmox.get_qemu_vms(
|
vms = proxmox.get_qemu_vms(node, vmid=vmid) + proxmox.get_lxc_vms(
|
||||||
vms_cluster_resources, vmid, node
|
node, vmid=vmid
|
||||||
) + proxmox.get_lxc_vms(vms_cluster_resources, vmid, node)
|
)
|
||||||
|
|
||||||
if vms or vmid is None:
|
if vms or vmid is None:
|
||||||
result["proxmox_vms"] = vms
|
result["proxmox_vms"] = vms
|
||||||
module.exit_json(**result)
|
module.exit_json(**result)
|
||||||
else:
|
else:
|
||||||
if node is None:
|
result["msg"] = "VM with vmid %s doesn't exist on node %s" % (vmid, node)
|
||||||
result["msg"] = "VM with vmid %s doesn't exist in cluster" % (vmid)
|
|
||||||
else:
|
|
||||||
result["msg"] = "VM with vmid %s doesn't exist on node %s" % (vmid, node)
|
|
||||||
module.fail_json(**result)
|
module.fail_json(**result)
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -28,276 +28,107 @@ from ansible_collections.community.general.tests.unit.plugins.modules.utils impo
|
||||||
)
|
)
|
||||||
import ansible_collections.community.general.plugins.module_utils.proxmox as proxmox_utils
|
import ansible_collections.community.general.plugins.module_utils.proxmox as proxmox_utils
|
||||||
|
|
||||||
NODE1 = "pve"
|
NODE = "pve"
|
||||||
NODE2 = "pve2"
|
LXC_VMS = [
|
||||||
RAW_CLUSTER_OUTPUT = [
|
|
||||||
{
|
{
|
||||||
"cpu": 0.174069059487628,
|
"uptime": 47,
|
||||||
"disk": 0,
|
|
||||||
"diskread": 6656,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "qemu/100",
|
|
||||||
"maxcpu": 1,
|
|
||||||
"maxdisk": 34359738368,
|
|
||||||
"maxmem": 4294967296,
|
|
||||||
"mem": 35304543,
|
|
||||||
"name": "pxe.home.arpa",
|
|
||||||
"netin": 416956,
|
|
||||||
"netout": 17330,
|
|
||||||
"node": NODE1,
|
|
||||||
"status": "running",
|
|
||||||
"template": 0,
|
|
||||||
"type": "qemu",
|
|
||||||
"uptime": 669,
|
|
||||||
"vmid": 100,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "qemu/101",
|
|
||||||
"maxcpu": 1,
|
|
||||||
"maxdisk": 0,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"mem": 0,
|
|
||||||
"name": "test1",
|
|
||||||
"netin": 0,
|
|
||||||
"netout": 0,
|
|
||||||
"node": NODE2,
|
|
||||||
"pool": "pool1",
|
|
||||||
"status": "stopped",
|
|
||||||
"template": 0,
|
|
||||||
"type": "qemu",
|
|
||||||
"uptime": 0,
|
|
||||||
"vmid": 101,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"disk": 352190464,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "lxc/102",
|
|
||||||
"maxcpu": 2,
|
|
||||||
"maxdisk": 10737418240,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"mem": 28192768,
|
|
||||||
"name": "test-lxc.home.arpa",
|
|
||||||
"netin": 102757,
|
|
||||||
"netout": 446,
|
|
||||||
"node": NODE1,
|
|
||||||
"status": "running",
|
|
||||||
"template": 0,
|
|
||||||
"type": "lxc",
|
|
||||||
"uptime": 161,
|
|
||||||
"vmid": 102,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "lxc/103",
|
|
||||||
"maxcpu": 2,
|
|
||||||
"maxdisk": 10737418240,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"mem": 0,
|
|
||||||
"name": "test1-lxc.home.arpa",
|
|
||||||
"netin": 0,
|
|
||||||
"netout": 0,
|
|
||||||
"node": NODE2,
|
|
||||||
"pool": "pool1",
|
|
||||||
"status": "stopped",
|
|
||||||
"template": 0,
|
|
||||||
"type": "lxc",
|
|
||||||
"uptime": 0,
|
|
||||||
"vmid": 103,
|
|
||||||
},
|
|
||||||
]
|
|
||||||
RAW_LXC_OUTPUT = [
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"cpus": 2,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"maxdisk": 10737418240,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"maxswap": 536870912,
|
"maxswap": 536870912,
|
||||||
"mem": 0,
|
|
||||||
"name": "test1-lxc.home.arpa",
|
|
||||||
"netin": 0,
|
|
||||||
"netout": 0,
|
|
||||||
"status": "stopped",
|
|
||||||
"swap": 0,
|
|
||||||
"type": "lxc",
|
|
||||||
"uptime": 0,
|
|
||||||
"vmid": "103",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"cpus": 2,
|
|
||||||
"disk": 352190464,
|
|
||||||
"diskread": 0,
|
"diskread": 0,
|
||||||
"diskwrite": 0,
|
|
||||||
"maxdisk": 10737418240,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"maxswap": 536870912,
|
|
||||||
"mem": 28192768,
|
|
||||||
"name": "test-lxc.home.arpa",
|
"name": "test-lxc.home.arpa",
|
||||||
"netin": 102757,
|
|
||||||
"netout": 446,
|
|
||||||
"pid": 4076752,
|
|
||||||
"status": "running",
|
"status": "running",
|
||||||
"swap": 0,
|
|
||||||
"type": "lxc",
|
|
||||||
"uptime": 161,
|
|
||||||
"vmid": "102",
|
"vmid": "102",
|
||||||
},
|
"type": "lxc",
|
||||||
]
|
"swap": 0,
|
||||||
RAW_QEMU_OUTPUT = [
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"cpus": 1,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"maxdisk": 0,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"mem": 0,
|
|
||||||
"name": "test1",
|
|
||||||
"netin": 0,
|
|
||||||
"netout": 0,
|
|
||||||
"status": "stopped",
|
|
||||||
"uptime": 0,
|
|
||||||
"vmid": 101,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0.174069059487628,
|
|
||||||
"cpus": 1,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 6656,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"maxdisk": 34359738368,
|
|
||||||
"maxmem": 4294967296,
|
|
||||||
"mem": 35304543,
|
|
||||||
"name": "pxe.home.arpa",
|
|
||||||
"netin": 416956,
|
|
||||||
"netout": 17330,
|
|
||||||
"pid": 4076688,
|
|
||||||
"status": "running",
|
|
||||||
"uptime": 669,
|
|
||||||
"vmid": 100,
|
|
||||||
},
|
|
||||||
]
|
|
||||||
EXPECTED_VMS_OUTPUT = [
|
|
||||||
{
|
|
||||||
"cpu": 0.174069059487628,
|
|
||||||
"cpus": 1,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 6656,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "qemu/100",
|
|
||||||
"maxcpu": 1,
|
|
||||||
"maxdisk": 34359738368,
|
|
||||||
"maxmem": 4294967296,
|
|
||||||
"mem": 35304543,
|
|
||||||
"name": "pxe.home.arpa",
|
|
||||||
"netin": 416956,
|
|
||||||
"netout": 17330,
|
|
||||||
"node": NODE1,
|
|
||||||
"pid": 4076688,
|
|
||||||
"status": "running",
|
|
||||||
"template": False,
|
|
||||||
"type": "qemu",
|
|
||||||
"uptime": 669,
|
|
||||||
"vmid": 100,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"cpus": 1,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "qemu/101",
|
|
||||||
"maxcpu": 1,
|
|
||||||
"maxdisk": 0,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"mem": 0,
|
|
||||||
"name": "test1",
|
|
||||||
"netin": 0,
|
|
||||||
"netout": 0,
|
|
||||||
"node": NODE2,
|
|
||||||
"pool": "pool1",
|
|
||||||
"status": "stopped",
|
|
||||||
"template": False,
|
|
||||||
"type": "qemu",
|
|
||||||
"uptime": 0,
|
|
||||||
"vmid": 101,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"cpu": 0,
|
|
||||||
"cpus": 2,
|
"cpus": 2,
|
||||||
"disk": 352190464,
|
"mem": 29134848,
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "lxc/102",
|
|
||||||
"maxcpu": 2,
|
|
||||||
"maxdisk": 10737418240,
|
"maxdisk": 10737418240,
|
||||||
"maxmem": 536870912,
|
"diskwrite": 0,
|
||||||
"maxswap": 536870912,
|
"netin": 35729,
|
||||||
"mem": 28192768,
|
|
||||||
"name": "test-lxc.home.arpa",
|
|
||||||
"netin": 102757,
|
|
||||||
"netout": 446,
|
"netout": 446,
|
||||||
"node": NODE1,
|
"pid": 1412780,
|
||||||
"pid": 4076752,
|
"maxmem": 536870912,
|
||||||
"status": "running",
|
"disk": 307625984,
|
||||||
"swap": 0,
|
"cpu": 0,
|
||||||
"template": False,
|
|
||||||
"type": "lxc",
|
|
||||||
"uptime": 161,
|
|
||||||
"vmid": 102,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cpu": 0,
|
|
||||||
"cpus": 2,
|
|
||||||
"disk": 0,
|
|
||||||
"diskread": 0,
|
|
||||||
"diskwrite": 0,
|
|
||||||
"id": "lxc/103",
|
|
||||||
"maxcpu": 2,
|
|
||||||
"maxdisk": 10737418240,
|
|
||||||
"maxmem": 536870912,
|
|
||||||
"maxswap": 536870912,
|
|
||||||
"mem": 0,
|
|
||||||
"name": "test1-lxc.home.arpa",
|
|
||||||
"netin": 0,
|
"netin": 0,
|
||||||
"netout": 0,
|
"netout": 0,
|
||||||
"node": NODE2,
|
"cpu": 0,
|
||||||
"pool": "pool1",
|
"maxmem": 536870912,
|
||||||
|
"disk": 0,
|
||||||
|
"name": "test1-lxc.home.arpa",
|
||||||
|
"diskread": 0,
|
||||||
"status": "stopped",
|
"status": "stopped",
|
||||||
"swap": 0,
|
"vmid": "103",
|
||||||
"template": False,
|
|
||||||
"type": "lxc",
|
"type": "lxc",
|
||||||
|
"swap": 0,
|
||||||
"uptime": 0,
|
"uptime": 0,
|
||||||
"vmid": 103,
|
"maxswap": 536870912,
|
||||||
|
"diskwrite": 0,
|
||||||
|
"cpus": 2,
|
||||||
|
"mem": 0,
|
||||||
|
"maxdisk": 10737418240,
|
||||||
|
},
|
||||||
|
]
|
||||||
|
QEMU_VMS = [
|
||||||
|
{
|
||||||
|
"vmid": 101,
|
||||||
|
"diskread": 0,
|
||||||
|
"status": "stopped",
|
||||||
|
"name": "test1",
|
||||||
|
"uptime": 0,
|
||||||
|
"diskwrite": 0,
|
||||||
|
"cpus": 1,
|
||||||
|
"mem": 0,
|
||||||
|
"maxdisk": 0,
|
||||||
|
"netout": 0,
|
||||||
|
"netin": 0,
|
||||||
|
"cpu": 0,
|
||||||
|
"maxmem": 536870912,
|
||||||
|
"disk": 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"netout": 4113,
|
||||||
|
"netin": 22738,
|
||||||
|
"pid": 1947197,
|
||||||
|
"maxmem": 4294967296,
|
||||||
|
"disk": 0,
|
||||||
|
"cpu": 0.0795350949559682,
|
||||||
|
"uptime": 41,
|
||||||
|
"vmid": 100,
|
||||||
|
"status": "running",
|
||||||
|
"diskread": 0,
|
||||||
|
"name": "pxe.home.arpa",
|
||||||
|
"cpus": 1,
|
||||||
|
"mem": 35315629,
|
||||||
|
"maxdisk": 34359738368,
|
||||||
|
"diskwrite": 0,
|
||||||
},
|
},
|
||||||
]
|
]
|
||||||
|
|
||||||
|
|
||||||
def get_module_args(type="all", node=None, vmid=None, name=None):
|
def get_module_args(type="all", vmid=None, name=None):
|
||||||
return {
|
return {
|
||||||
"api_host": "host",
|
"api_host": "host",
|
||||||
"api_user": "user",
|
"api_user": "user",
|
||||||
"api_password": "password",
|
"api_password": "password",
|
||||||
"node": node,
|
"node": NODE,
|
||||||
"type": type,
|
"type": type,
|
||||||
"vmid": vmid,
|
"vmid": vmid,
|
||||||
"name": name,
|
"name": name,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def normalized_expected_vms_output(vms):
|
||||||
|
result = [vm.copy() for vm in vms]
|
||||||
|
for vm in result:
|
||||||
|
if "type" not in vm:
|
||||||
|
# response for QEMU VMs doesn't contain type field, adding it
|
||||||
|
vm["type"] = "qemu"
|
||||||
|
vm["vmid"] = int(vm["vmid"])
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
class TestProxmoxVmInfoModule(ModuleTestCase):
|
class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
def setUp(self):
|
def setUp(self):
|
||||||
super(TestProxmoxVmInfoModule, self).setUp()
|
super(TestProxmoxVmInfoModule, self).setUp()
|
||||||
|
@ -307,15 +138,12 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
"ansible_collections.community.general.plugins.module_utils.proxmox.ProxmoxAnsible._connect",
|
"ansible_collections.community.general.plugins.module_utils.proxmox.ProxmoxAnsible._connect",
|
||||||
).start()
|
).start()
|
||||||
self.connect_mock.return_value.nodes.return_value.lxc.return_value.get.return_value = (
|
self.connect_mock.return_value.nodes.return_value.lxc.return_value.get.return_value = (
|
||||||
RAW_LXC_OUTPUT
|
LXC_VMS
|
||||||
)
|
)
|
||||||
self.connect_mock.return_value.nodes.return_value.qemu.return_value.get.return_value = (
|
self.connect_mock.return_value.nodes.return_value.qemu.return_value.get.return_value = (
|
||||||
RAW_QEMU_OUTPUT
|
QEMU_VMS
|
||||||
)
|
)
|
||||||
self.connect_mock.return_value.cluster.return_value.resources.return_value.get.return_value = (
|
self.connect_mock.return_value.nodes.get.return_value = [{"node": NODE}]
|
||||||
RAW_CLUSTER_OUTPUT
|
|
||||||
)
|
|
||||||
self.connect_mock.return_value.nodes.get.return_value = [{"node": NODE1}]
|
|
||||||
|
|
||||||
def tearDown(self):
|
def tearDown(self):
|
||||||
self.connect_mock.stop()
|
self.connect_mock.stop()
|
||||||
|
@ -327,7 +155,7 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
assert result["msg"] == "missing required arguments: api_host, api_user"
|
assert result["msg"] == "missing required arguments: api_host, api_user, node"
|
||||||
|
|
||||||
def test_get_lxc_vms_information(self):
|
def test_get_lxc_vms_information(self):
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
with pytest.raises(AnsibleExitJson) as exc_info:
|
||||||
|
@ -336,34 +164,36 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
assert result["changed"] is False
|
assert result["changed"] is False
|
||||||
assert result["proxmox_vms"] == [
|
assert result["proxmox_vms"] == LXC_VMS
|
||||||
vm for vm in EXPECTED_VMS_OUTPUT if vm["type"] == "lxc"
|
|
||||||
]
|
|
||||||
|
|
||||||
def test_get_qemu_vms_information(self):
|
def test_get_qemu_vms_information(self):
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
with pytest.raises(AnsibleExitJson) as exc_info:
|
||||||
|
expected_output = normalized_expected_vms_output(QEMU_VMS)
|
||||||
set_module_args(get_module_args(type="qemu"))
|
set_module_args(get_module_args(type="qemu"))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
assert result["proxmox_vms"] == [
|
assert result["proxmox_vms"] == expected_output
|
||||||
vm for vm in EXPECTED_VMS_OUTPUT if vm["type"] == "qemu"
|
|
||||||
]
|
|
||||||
|
|
||||||
def test_get_all_vms_information(self):
|
def test_get_all_vms_information(self):
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
with pytest.raises(AnsibleExitJson) as exc_info:
|
||||||
|
qemu_output = normalized_expected_vms_output(QEMU_VMS)
|
||||||
|
expected_output = qemu_output + LXC_VMS
|
||||||
|
|
||||||
set_module_args(get_module_args())
|
set_module_args(get_module_args())
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
assert result["proxmox_vms"] == EXPECTED_VMS_OUTPUT
|
assert result["proxmox_vms"] == expected_output
|
||||||
|
|
||||||
def test_vmid_is_converted_to_int(self):
|
def test_vmid_is_converted_to_int(self):
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
with pytest.raises(AnsibleExitJson) as exc_info:
|
||||||
|
expected_output = normalized_expected_vms_output(LXC_VMS)
|
||||||
set_module_args(get_module_args(type="lxc"))
|
set_module_args(get_module_args(type="lxc"))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
|
assert result["proxmox_vms"] == expected_output
|
||||||
assert isinstance(result["proxmox_vms"][0]["vmid"], int)
|
assert isinstance(result["proxmox_vms"][0]["vmid"], int)
|
||||||
|
|
||||||
def test_get_specific_lxc_vm_information(self):
|
def test_get_specific_lxc_vm_information(self):
|
||||||
|
@ -371,8 +201,8 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
vmid = 102
|
vmid = 102
|
||||||
expected_output = [
|
expected_output = [
|
||||||
vm
|
vm
|
||||||
for vm in EXPECTED_VMS_OUTPUT
|
for vm in normalized_expected_vms_output(LXC_VMS)
|
||||||
if vm["vmid"] == vmid and vm["type"] == "lxc"
|
if vm["vmid"] == vmid
|
||||||
]
|
]
|
||||||
set_module_args(get_module_args(type="lxc", vmid=vmid))
|
set_module_args(get_module_args(type="lxc", vmid=vmid))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
@ -386,8 +216,8 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
vmid = 100
|
vmid = 100
|
||||||
expected_output = [
|
expected_output = [
|
||||||
vm
|
vm
|
||||||
for vm in EXPECTED_VMS_OUTPUT
|
for vm in normalized_expected_vms_output(QEMU_VMS)
|
||||||
if vm["vmid"] == vmid and vm["type"] == "qemu"
|
if vm["vmid"] == vmid
|
||||||
]
|
]
|
||||||
set_module_args(get_module_args(type="qemu", vmid=vmid))
|
set_module_args(get_module_args(type="qemu", vmid=vmid))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
@ -399,7 +229,11 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
def test_get_specific_vm_information(self):
|
def test_get_specific_vm_information(self):
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
with pytest.raises(AnsibleExitJson) as exc_info:
|
||||||
vmid = 100
|
vmid = 100
|
||||||
expected_output = [vm for vm in EXPECTED_VMS_OUTPUT if vm["vmid"] == vmid]
|
expected_output = [
|
||||||
|
vm
|
||||||
|
for vm in normalized_expected_vms_output(QEMU_VMS + LXC_VMS)
|
||||||
|
if vm["vmid"] == vmid
|
||||||
|
]
|
||||||
set_module_args(get_module_args(type="all", vmid=vmid))
|
set_module_args(get_module_args(type="all", vmid=vmid))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
|
@ -414,7 +248,11 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
]
|
]
|
||||||
|
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
with pytest.raises(AnsibleExitJson) as exc_info:
|
||||||
expected_output = [vm for vm in EXPECTED_VMS_OUTPUT if vm["name"] == name]
|
expected_output = [
|
||||||
|
vm
|
||||||
|
for vm in normalized_expected_vms_output(QEMU_VMS + LXC_VMS)
|
||||||
|
if vm["name"] == name
|
||||||
|
]
|
||||||
set_module_args(get_module_args(type="all", name=name))
|
set_module_args(get_module_args(type="all", name=name))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
|
@ -422,65 +260,14 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
assert result["proxmox_vms"] == expected_output
|
assert result["proxmox_vms"] == expected_output
|
||||||
assert len(result["proxmox_vms"]) == 1
|
assert len(result["proxmox_vms"]) == 1
|
||||||
|
|
||||||
def test_get_all_lxc_vms_from_specific_node(self):
|
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
|
||||||
expected_output = [
|
|
||||||
vm
|
|
||||||
for vm in EXPECTED_VMS_OUTPUT
|
|
||||||
if vm["node"] == NODE1 and vm["type"] == "lxc"
|
|
||||||
]
|
|
||||||
set_module_args(get_module_args(type="lxc", node=NODE1))
|
|
||||||
self.module.main()
|
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
|
||||||
assert result["proxmox_vms"] == expected_output
|
|
||||||
assert len(result["proxmox_vms"]) == 1
|
|
||||||
|
|
||||||
def test_get_all_qemu_vms_from_specific_node(self):
|
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
|
||||||
expected_output = [
|
|
||||||
vm
|
|
||||||
for vm in EXPECTED_VMS_OUTPUT
|
|
||||||
if vm["node"] == NODE1 and vm["type"] == "qemu"
|
|
||||||
]
|
|
||||||
set_module_args(get_module_args(type="qemu", node=NODE1))
|
|
||||||
self.module.main()
|
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
|
||||||
assert result["proxmox_vms"] == expected_output
|
|
||||||
assert len(result["proxmox_vms"]) == 1
|
|
||||||
|
|
||||||
def test_get_all_vms_from_specific_node(self):
|
|
||||||
with pytest.raises(AnsibleExitJson) as exc_info:
|
|
||||||
expected_output = [
|
|
||||||
vm
|
|
||||||
for vm in EXPECTED_VMS_OUTPUT
|
|
||||||
if vm["node"] == NODE1
|
|
||||||
]
|
|
||||||
set_module_args(get_module_args(node=NODE1))
|
|
||||||
self.module.main()
|
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
|
||||||
assert result["proxmox_vms"] == expected_output
|
|
||||||
assert len(result["proxmox_vms"]) == 2
|
|
||||||
|
|
||||||
def test_module_fail_when_vm_does_not_exist_on_node(self):
|
def test_module_fail_when_vm_does_not_exist_on_node(self):
|
||||||
with pytest.raises(AnsibleFailJson) as exc_info:
|
|
||||||
vmid = 200
|
|
||||||
set_module_args(get_module_args(type="all", vmid=vmid, node=NODE1))
|
|
||||||
self.module.main()
|
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
|
||||||
assert result["msg"] == "VM with vmid 200 doesn't exist on node pve"
|
|
||||||
|
|
||||||
def test_module_fail_when_vm_does_not_exist_in_cluster(self):
|
|
||||||
with pytest.raises(AnsibleFailJson) as exc_info:
|
with pytest.raises(AnsibleFailJson) as exc_info:
|
||||||
vmid = 200
|
vmid = 200
|
||||||
set_module_args(get_module_args(type="all", vmid=vmid))
|
set_module_args(get_module_args(type="all", vmid=vmid))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
assert result["msg"] == "VM with vmid 200 doesn't exist in cluster"
|
assert result["msg"] == "VM with vmid 200 doesn't exist on node pve"
|
||||||
|
|
||||||
def test_module_fail_when_qemu_request_fails(self):
|
def test_module_fail_when_qemu_request_fails(self):
|
||||||
self.connect_mock.return_value.nodes.return_value.qemu.return_value.get.side_effect = IOError(
|
self.connect_mock.return_value.nodes.return_value.qemu.return_value.get.side_effect = IOError(
|
||||||
|
@ -504,24 +291,10 @@ class TestProxmoxVmInfoModule(ModuleTestCase):
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
assert "Failed to retrieve LXC VMs information:" in result["msg"]
|
assert "Failed to retrieve LXC VMs information:" in result["msg"]
|
||||||
|
|
||||||
def test_module_fail_when_cluster_resources_request_fails(self):
|
|
||||||
self.connect_mock.return_value.cluster.return_value.resources.return_value.get.side_effect = IOError(
|
|
||||||
"Some mocked connection error."
|
|
||||||
)
|
|
||||||
with pytest.raises(AnsibleFailJson) as exc_info:
|
|
||||||
set_module_args(get_module_args())
|
|
||||||
self.module.main()
|
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
|
||||||
assert (
|
|
||||||
"Failed to retrieve VMs information from cluster resources:"
|
|
||||||
in result["msg"]
|
|
||||||
)
|
|
||||||
|
|
||||||
def test_module_fail_when_node_does_not_exist(self):
|
def test_module_fail_when_node_does_not_exist(self):
|
||||||
self.connect_mock.return_value.nodes.get.return_value = []
|
self.connect_mock.return_value.nodes.get.return_value = []
|
||||||
with pytest.raises(AnsibleFailJson) as exc_info:
|
with pytest.raises(AnsibleFailJson) as exc_info:
|
||||||
set_module_args(get_module_args(type="all", node=NODE1))
|
set_module_args(get_module_args(type="all"))
|
||||||
self.module.main()
|
self.module.main()
|
||||||
|
|
||||||
result = exc_info.value.args[0]
|
result = exc_info.value.args[0]
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue