mirror of
https://github.com/ansible-collections/community.general.git
synced 2025-05-29 20:39:09 -07:00
* add support to vmware_guest for template => vm conversion While the vmware_guest currently supports conversion of VMs to templates using the is_template argument, it does not support the inverse: converting templates back into VMs. This change adds that functionality. When converting a template back into a VM, the extra config option "uuid.action" is also set so that VMware will automatically create a new UUID for the converted VM. If the "uuid.action" setting is already configured, it will not be modified. Setting this prevents an interactive question from being raised when attempting to boot the VM. * Add integration tests for vmware_guest is_template * Add additional idempotency test for vmware_guest is_template
1571 lines
69 KiB
Python
1571 lines
69 KiB
Python
#!/usr/bin/python
|
|
# -*- coding: utf-8 -*-
|
|
#
|
|
# This module is also sponsored by E.T.A.I. (www.etai.fr)
|
|
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
|
|
from __future__ import absolute_import, division, print_function
|
|
__metaclass__ = type
|
|
|
|
|
|
ANSIBLE_METADATA = {'metadata_version': '1.1',
|
|
'status': ['preview'],
|
|
'supported_by': 'community'}
|
|
|
|
|
|
DOCUMENTATION = r'''
|
|
---
|
|
module: vmware_guest
|
|
short_description: Manages virtual machines in vCenter
|
|
description:
|
|
- Create new virtual machines (from templates or not).
|
|
- Power on/power off/restart a virtual machine.
|
|
- Modify, rename or remove a virtual machine.
|
|
version_added: '2.2'
|
|
author:
|
|
- James Tanner (@jctanner) <tanner.jc@gmail.com>
|
|
- Loic Blot (@nerzhul) <loic.blot@unix-experience.fr>
|
|
- Philippe Dellaert (@pdellaert) <philippe@dellaert.org>
|
|
notes:
|
|
- Tested on vSphere 5.5 and 6.0
|
|
requirements:
|
|
- python >= 2.6
|
|
- PyVmomi
|
|
options:
|
|
state:
|
|
description:
|
|
- What state should the virtual machine be in?
|
|
- If C(state) is set to C(present) and VM exists, ensure the VM configuration conforms to task arguments.
|
|
required: yes
|
|
choices: [ 'present', 'absent', 'poweredon', 'poweredoff', 'restarted', 'suspended', 'shutdownguest', 'rebootguest' ]
|
|
name:
|
|
description:
|
|
- Name of the VM to work with.
|
|
- VM names in vCenter are not necessarily unique, which may be problematic, see C(name_match).
|
|
required: yes
|
|
name_match:
|
|
description:
|
|
- If multiple VMs matching the name, use the first or last found.
|
|
default: 'first'
|
|
choices: [ 'first', 'last' ]
|
|
uuid:
|
|
description:
|
|
- UUID of the instance to manage if known, this is VMware's unique identifier.
|
|
- This is required if name is not supplied.
|
|
template:
|
|
description:
|
|
- Template used to create VM.
|
|
- If this value is not set, VM is created without using a template.
|
|
- If the VM exists already this setting will be ignored.
|
|
is_template:
|
|
description:
|
|
- Flag the instance as a template.
|
|
default: 'no'
|
|
type: bool
|
|
version_added: '2.3'
|
|
folder:
|
|
description:
|
|
- Destination folder, absolute path to find an existing guest or create the new guest.
|
|
- The folder should include the datacenter. ESX's datacenter is ha-datacenter
|
|
- 'Examples:'
|
|
- ' folder: /ha-datacenter/vm'
|
|
- ' folder: ha-datacenter/vm'
|
|
- ' folder: /datacenter1/vm'
|
|
- ' folder: datacenter1/vm'
|
|
- ' folder: /datacenter1/vm/folder1'
|
|
- ' folder: datacenter1/vm/folder1'
|
|
- ' folder: /folder1/datacenter1/vm'
|
|
- ' folder: folder1/datacenter1/vm'
|
|
- ' folder: /folder1/datacenter1/vm/folder2'
|
|
default: /vm
|
|
hardware:
|
|
description:
|
|
- Manage some VM hardware attributes.
|
|
- 'Valid attributes are:'
|
|
- ' - C(memory_mb) (integer): Amount of memory in MB.'
|
|
- ' - C(num_cpus) (integer): Number of CPUs.'
|
|
- ' - C(scsi) (string): Valid values are C(buslogic), C(lsilogic), C(lsilogicsas) and C(paravirtual) (default).'
|
|
guest_id:
|
|
description:
|
|
- Set the guest ID (Debian, RHEL, Windows...).
|
|
- This field is required when creating a VM.
|
|
- >
|
|
Valid values are referenced here:
|
|
https://www.vmware.com/support/developer/converter-sdk/conv55_apireference/vim.vm.GuestOsDescriptor.GuestOsIdentifier.html
|
|
version_added: '2.3'
|
|
disk:
|
|
description:
|
|
- A list of disks to add.
|
|
- 'Valid attributes are:'
|
|
- ' - C(size_[tb,gb,mb,kb]) (integer): Disk storage size in specified unit.'
|
|
- ' - C(type) (string): Valid value is C(thin) (default: None).'
|
|
- ' - C(datastore) (string): Datastore to use for the disk. If C(autoselect_datastore) is enabled, filter datastore selection.'
|
|
- ' - C(autoselect_datastore) (bool): select the less used datastore.'
|
|
cdrom:
|
|
description:
|
|
- A CD-ROM configuration for the VM.
|
|
- 'Valid attributes are:'
|
|
- ' - C(type) (string): The type of CD-ROM, valid options are C(none), C(client) or C(iso). With C(none) the CD-ROM will be disconnected but present.'
|
|
- ' - C(iso_path) (string): The datastore path to the ISO file to use, in the form of C([datastore1] path/to/file.iso). Required if type is iso.'
|
|
version_added: '2.5'
|
|
resource_pool:
|
|
description:
|
|
- Affect machine to the given resource pool.
|
|
- Resource pool should be child of the selected host parent.
|
|
version_added: '2.3'
|
|
wait_for_ip_address:
|
|
description:
|
|
- Wait until vCenter detects an IP address for the VM.
|
|
- This requires vmware-tools (vmtoolsd) to properly work after creation.
|
|
default: 'no'
|
|
type: bool
|
|
snapshot_src:
|
|
description:
|
|
- Name of an existing snapshot to use to create a clone of a VM.
|
|
version_added: '2.4'
|
|
linked_clone:
|
|
description:
|
|
- Whether to create a Linked Clone from the snapshot specified.
|
|
default: 'no'
|
|
type: bool
|
|
version_added: '2.4'
|
|
force:
|
|
description:
|
|
- Ignore warnings and complete the actions.
|
|
default: 'no'
|
|
type: bool
|
|
datacenter:
|
|
description:
|
|
- Destination datacenter for the deploy operation.
|
|
default: ha-datacenter
|
|
cluster:
|
|
description:
|
|
- The cluster name where the VM will run.
|
|
version_added: '2.3'
|
|
esxi_hostname:
|
|
description:
|
|
- The ESXi hostname where the VM will run.
|
|
annotation:
|
|
description:
|
|
- A note or annotation to include in the VM.
|
|
version_added: '2.3'
|
|
customvalues:
|
|
description:
|
|
- Define a list of customvalues to set on VM.
|
|
- A customvalue object takes 2 fields C(key) and C(value).
|
|
version_added: '2.3'
|
|
networks:
|
|
description:
|
|
- A list of networks (in the order of the NICs).
|
|
- 'One of the below parameters is required per entry:'
|
|
- ' - C(name) (string): Name of the portgroup for this interface.'
|
|
- ' - C(vlan) (integer): VLAN number for this interface.'
|
|
- 'Optional parameters per entry (used for virtual hardware):'
|
|
- ' - C(device_type) (string): Virtual network device (one of C(e1000), C(e1000e), C(pcnet32), C(vmxnet2), C(vmxnet3) (default), C(sriov)).'
|
|
- ' - C(mac) (string): Customize mac address.'
|
|
- 'Optional parameters per entry (used for OS customization):'
|
|
- ' - C(type) (string): Type of IP assignment (either C(dhcp) or C(static)).'
|
|
- ' - C(ip) (string): Static IP address (implies C(type: static)).'
|
|
- ' - C(netmask) (string): Static netmask required for C(ip).'
|
|
- ' - C(gateway) (string): Static gateway.'
|
|
- ' - C(dns_servers) (string): DNS servers for this network interface (Windows).'
|
|
- ' - C(domain) (string): Domain name for this network interface (Windows).'
|
|
version_added: '2.3'
|
|
customization:
|
|
description:
|
|
- Parameters for OS customization when cloning from template.
|
|
- 'Common parameters (Linux/Windows):'
|
|
- ' - C(dns_servers) (list): List of DNS servers to configure.'
|
|
- ' - C(dns_suffix) (list): List of domain suffixes, aka DNS search path (default: C(domain) parameter).'
|
|
- ' - C(domain) (string): DNS domain name to use.'
|
|
- ' - C(hostname) (string): Computer hostname (default: shorted C(name) parameter).'
|
|
- 'Parameters related to Windows customization:'
|
|
- ' - C(autologon) (bool): Auto logon after VM customization (default: False).'
|
|
- ' - C(autologoncount) (int): Number of autologon after reboot (default: 1).'
|
|
- ' - C(domainadmin) (string): User used to join in AD domain (mandatory with C(joindomain)).'
|
|
- ' - C(domainadminpassword) (string): Password used to join in AD domain (mandatory with C(joindomain)).'
|
|
- ' - C(fullname) (string): Server owner name (default: Administrator).'
|
|
- ' - C(joindomain) (string): AD domain to join (Not compatible with C(joinworkgroup)).'
|
|
- ' - C(joinworkgroup) (string): Workgroup to join (Not compatible with C(joindomain), default: WORKGROUP).'
|
|
- ' - C(orgname) (string): Organisation name (default: ACME).'
|
|
- ' - C(password) (string): Local administrator password.'
|
|
- ' - C(productid) (string): Product ID.'
|
|
- ' - C(runonce) (list): List of commands to run at first user logon.'
|
|
- ' - C(timezone) (int): Timezone (See U(https://msdn.microsoft.com/en-us/library/ms912391.aspx)).'
|
|
version_added: '2.3'
|
|
extends_documentation_fragment: vmware.documentation
|
|
'''
|
|
|
|
EXAMPLES = r'''
|
|
- name: Create a VM from a template
|
|
vmware_guest:
|
|
hostname: 192.0.2.44
|
|
username: administrator@vsphere.local
|
|
password: vmware
|
|
validate_certs: no
|
|
folder: /testvms
|
|
name: testvm_2
|
|
state: poweredon
|
|
template: template_el7
|
|
disk:
|
|
- size_gb: 10
|
|
type: thin
|
|
datastore: g73_datastore
|
|
hardware:
|
|
memory_mb: 512
|
|
num_cpus: 1
|
|
scsi: paravirtual
|
|
cdrom:
|
|
type: iso
|
|
iso_path: "[datastore1] livecd.iso"
|
|
networks:
|
|
- name: VM Network
|
|
mac: aa:bb:dd:aa:00:14
|
|
wait_for_ip_address: yes
|
|
delegate_to: localhost
|
|
register: deploy
|
|
|
|
- name: Clone a VM from Template and customize
|
|
vmware_guest:
|
|
hostname: 192.168.1.209
|
|
username: administrator@vsphere.local
|
|
password: vmware
|
|
validate_certs: no
|
|
datacenter: datacenter1
|
|
cluster: cluster
|
|
name: testvm-2
|
|
template: template_windows
|
|
networks:
|
|
- name: VM Network
|
|
ip: 192.168.1.100
|
|
netmask: 255.255.255.0
|
|
gateway: 192.168.1.1
|
|
mac: aa:bb:dd:aa:00:14
|
|
domain: my_domain
|
|
dns_servers:
|
|
- 192.168.1.1
|
|
- 192.168.1.2
|
|
- vlan: 1234
|
|
type: dhcp
|
|
customization:
|
|
autologon: yes
|
|
dns_servers:
|
|
- 192.168.1.1
|
|
- 192.168.1.2
|
|
domain: my_domain
|
|
password: new_vm_password
|
|
runonce:
|
|
- powershell.exe -ExecutionPolicy Unrestricted -File C:\Windows\Temp\ConfigureRemotingForAnsible.ps1 -ForceNewSSLCert -EnableCredSSP
|
|
delegate_to: localhost
|
|
|
|
- name: Create a VM template
|
|
vmware_guest:
|
|
hostname: 192.0.2.88
|
|
username: administrator@vsphere.local
|
|
password: vmware
|
|
validate_certs: no
|
|
datacenter: datacenter1
|
|
cluster: vmware_cluster_esx
|
|
resource_pool: highperformance_pool
|
|
folder: /testvms
|
|
name: testvm_6
|
|
is_template: yes
|
|
guest_id: debian6_64Guest
|
|
disk:
|
|
- size_gb: 10
|
|
type: thin
|
|
datastore: g73_datastore
|
|
hardware:
|
|
memory_mb: 512
|
|
num_cpus: 1
|
|
scsi: lsilogic
|
|
delegate_to: localhost
|
|
register: deploy
|
|
|
|
- name: Rename a VM (requires the VM's uuid)
|
|
vmware_guest:
|
|
hostname: 192.168.1.209
|
|
username: administrator@vsphere.local
|
|
password: vmware
|
|
uuid: 421e4592-c069-924d-ce20-7e7533fab926
|
|
name: new_name
|
|
state: present
|
|
delegate_to: localhost
|
|
|
|
- name: Remove a VM by uuid
|
|
vmware_guest:
|
|
hostname: 192.168.1.209
|
|
username: administrator@vsphere.local
|
|
password: vmware
|
|
uuid: 421e4592-c069-924d-ce20-7e7533fab926
|
|
state: absent
|
|
delegate_to: localhost
|
|
'''
|
|
|
|
RETURN = r'''
|
|
instance:
|
|
description: metadata about the new virtualmachine
|
|
returned: always
|
|
type: dict
|
|
sample: None
|
|
'''
|
|
|
|
import time
|
|
|
|
HAS_PYVMOMI = False
|
|
try:
|
|
import pyVmomi
|
|
from pyVmomi import vim
|
|
|
|
HAS_PYVMOMI = True
|
|
except ImportError:
|
|
pass
|
|
|
|
from ansible.module_utils.basic import AnsibleModule
|
|
from ansible.module_utils._text import to_text
|
|
from ansible.module_utils.vmware import (find_obj, gather_vm_facts, get_all_objs,
|
|
compile_folder_path_for_object, serialize_spec,
|
|
vmware_argument_spec, set_vm_power_state, PyVmomi)
|
|
|
|
|
|
class PyVmomiDeviceHelper(object):
|
|
""" This class is a helper to create easily VMWare Objects for PyVmomiHelper """
|
|
|
|
def __init__(self, module):
|
|
self.module = module
|
|
self.next_disk_unit_number = 0
|
|
|
|
@staticmethod
|
|
def create_scsi_controller(scsi_type):
|
|
scsi_ctl = vim.vm.device.VirtualDeviceSpec()
|
|
scsi_ctl.operation = vim.vm.device.VirtualDeviceSpec.Operation.add
|
|
if scsi_type == 'lsilogic':
|
|
scsi_ctl.device = vim.vm.device.VirtualLsiLogicController()
|
|
elif scsi_type == 'paravirtual':
|
|
scsi_ctl.device = vim.vm.device.ParaVirtualSCSIController()
|
|
elif scsi_type == 'buslogic':
|
|
scsi_ctl.device = vim.vm.device.VirtualBusLogicController()
|
|
elif scsi_type == 'lsilogicsas':
|
|
scsi_ctl.device = vim.vm.device.VirtualLsiLogicSASController()
|
|
|
|
scsi_ctl.device.deviceInfo = vim.Description()
|
|
scsi_ctl.device.slotInfo = vim.vm.device.VirtualDevice.PciBusSlotInfo()
|
|
scsi_ctl.device.slotInfo.pciSlotNumber = 16
|
|
scsi_ctl.device.controllerKey = 100
|
|
scsi_ctl.device.unitNumber = 3
|
|
scsi_ctl.device.busNumber = 0
|
|
scsi_ctl.device.hotAddRemove = True
|
|
scsi_ctl.device.sharedBus = 'noSharing'
|
|
scsi_ctl.device.scsiCtlrUnitNumber = 7
|
|
|
|
return scsi_ctl
|
|
|
|
@staticmethod
|
|
def is_scsi_controller(device):
|
|
return isinstance(device, vim.vm.device.VirtualLsiLogicController) or \
|
|
isinstance(device, vim.vm.device.ParaVirtualSCSIController) or \
|
|
isinstance(device, vim.vm.device.VirtualBusLogicController) or \
|
|
isinstance(device, vim.vm.device.VirtualLsiLogicSASController)
|
|
|
|
@staticmethod
|
|
def create_ide_controller():
|
|
ide_ctl = vim.vm.device.VirtualDeviceSpec()
|
|
ide_ctl.operation = vim.vm.device.VirtualDeviceSpec.Operation.add
|
|
ide_ctl.device = vim.vm.device.VirtualIDEController()
|
|
ide_ctl.device.deviceInfo = vim.Description()
|
|
ide_ctl.device.busNumber = 0
|
|
|
|
return ide_ctl
|
|
|
|
@staticmethod
|
|
def create_cdrom(ide_ctl, cdrom_type, iso_path=None):
|
|
cdrom_spec = vim.vm.device.VirtualDeviceSpec()
|
|
cdrom_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.add
|
|
cdrom_spec.device = vim.vm.device.VirtualCdrom()
|
|
cdrom_spec.device.controllerKey = ide_ctl.device.key
|
|
cdrom_spec.device.key = -1
|
|
cdrom_spec.device.connectable = vim.vm.device.VirtualDevice.ConnectInfo()
|
|
cdrom_spec.device.connectable.allowGuestControl = True
|
|
cdrom_spec.device.connectable.startConnected = (cdrom_type != "none")
|
|
if cdrom_type in ["none", "client"]:
|
|
cdrom_spec.device.backing = vim.vm.device.VirtualCdrom.RemotePassthroughBackingInfo()
|
|
elif cdrom_type == "iso":
|
|
cdrom_spec.device.backing = vim.vm.device.VirtualCdrom.IsoBackingInfo(fileName=iso_path)
|
|
|
|
return cdrom_spec
|
|
|
|
@staticmethod
|
|
def is_equal_cdrom(vm_obj, cdrom_device, cdrom_type, iso_path):
|
|
if cdrom_type == "none":
|
|
return (isinstance(cdrom_device.backing, vim.vm.device.VirtualCdrom.RemotePassthroughBackingInfo) and
|
|
cdrom_device.connectable.allowGuestControl and
|
|
not cdrom_device.connectable.startConnected and
|
|
(vm_obj.runtime.powerState != vim.VirtualMachinePowerState.poweredOn or not cdrom_device.connectable.connected))
|
|
elif cdrom_type == "client":
|
|
return (isinstance(cdrom_device.backing, vim.vm.device.VirtualCdrom.RemotePassthroughBackingInfo) and
|
|
cdrom_device.connectable.allowGuestControl and
|
|
cdrom_device.connectable.startConnected and
|
|
(vm_obj.runtime.powerState != vim.VirtualMachinePowerState.poweredOn or cdrom_device.connectable.connected))
|
|
elif cdrom_type == "iso":
|
|
return (isinstance(cdrom_device.backing, vim.vm.device.VirtualCdrom.IsoBackingInfo) and
|
|
cdrom_device.backing.fileName == iso_path and
|
|
cdrom_device.connectable.allowGuestControl and
|
|
cdrom_device.connectable.startConnected and
|
|
(vm_obj.runtime.powerState != vim.VirtualMachinePowerState.poweredOn or cdrom_device.connectable.connected))
|
|
|
|
def create_scsi_disk(self, scsi_ctl, disk_index=None):
|
|
diskspec = vim.vm.device.VirtualDeviceSpec()
|
|
diskspec.operation = vim.vm.device.VirtualDeviceSpec.Operation.add
|
|
diskspec.fileOperation = vim.vm.device.VirtualDeviceSpec.FileOperation.create
|
|
diskspec.device = vim.vm.device.VirtualDisk()
|
|
diskspec.device.backing = vim.vm.device.VirtualDisk.FlatVer2BackingInfo()
|
|
diskspec.device.backing.diskMode = 'persistent'
|
|
diskspec.device.controllerKey = scsi_ctl.device.key
|
|
|
|
assert self.next_disk_unit_number != 7
|
|
assert disk_index != 7
|
|
"""
|
|
Configure disk unit number.
|
|
"""
|
|
if disk_index is not None:
|
|
diskspec.device.unitNumber = disk_index
|
|
self.next_disk_unit_number = disk_index + 1
|
|
else:
|
|
diskspec.device.unitNumber = self.next_disk_unit_number
|
|
self.next_disk_unit_number += 1
|
|
|
|
# unit number 7 is reserved to SCSI controller, increase next index
|
|
if self.next_disk_unit_number == 7:
|
|
self.next_disk_unit_number += 1
|
|
|
|
return diskspec
|
|
|
|
def create_nic(self, device_type, device_label, device_infos):
|
|
nic = vim.vm.device.VirtualDeviceSpec()
|
|
if device_type == 'pcnet32':
|
|
nic.device = vim.vm.device.VirtualPCNet32()
|
|
elif device_type == 'vmxnet2':
|
|
nic.device = vim.vm.device.VirtualVmxnet2()
|
|
elif device_type == 'vmxnet3':
|
|
nic.device = vim.vm.device.VirtualVmxnet3()
|
|
elif device_type == 'e1000':
|
|
nic.device = vim.vm.device.VirtualE1000()
|
|
elif device_type == 'e1000e':
|
|
nic.device = vim.vm.device.VirtualE1000e()
|
|
elif device_type == 'sriov':
|
|
nic.device = vim.vm.device.VirtualSriovEthernetCard()
|
|
else:
|
|
self.module.fail_json(msg='Invalid device_type "%s" for network "%s"' % (device_type, device_infos['name']))
|
|
|
|
nic.device.wakeOnLanEnabled = True
|
|
nic.device.deviceInfo = vim.Description()
|
|
nic.device.deviceInfo.label = device_label
|
|
nic.device.deviceInfo.summary = device_infos['name']
|
|
nic.device.connectable = vim.vm.device.VirtualDevice.ConnectInfo()
|
|
nic.device.connectable.startConnected = True
|
|
nic.device.connectable.allowGuestControl = True
|
|
nic.device.connectable.connected = True
|
|
if 'mac' in device_infos:
|
|
nic.device.addressType = 'assigned'
|
|
nic.device.macAddress = device_infos['mac']
|
|
else:
|
|
nic.device.addressType = 'generated'
|
|
|
|
return nic
|
|
|
|
|
|
class PyVmomiCache(object):
|
|
""" This class caches references to objects which are requested multiples times but not modified """
|
|
def __init__(self, content, dc_name=None):
|
|
self.content = content
|
|
self.dc_name = dc_name
|
|
self.networks = {}
|
|
self.clusters = {}
|
|
self.esx_hosts = {}
|
|
self.parent_datacenters = {}
|
|
|
|
def find_obj(self, content, types, name, confine_to_datacenter=True):
|
|
""" Wrapper around find_obj to set datacenter context """
|
|
result = find_obj(content, types, name)
|
|
if result and confine_to_datacenter:
|
|
if self.get_parent_datacenter(result).name != self.dc_name:
|
|
result = None
|
|
objects = self.get_all_objs(content, types, confine_to_datacenter=True)
|
|
for obj in objects:
|
|
if name is None or obj.name == name:
|
|
return obj
|
|
return result
|
|
|
|
def get_all_objs(self, content, types, confine_to_datacenter=True):
|
|
""" Wrapper around get_all_objs to set datacenter context """
|
|
objects = get_all_objs(content, types)
|
|
if confine_to_datacenter:
|
|
if hasattr(objects, 'items'):
|
|
# resource pools come back as a dictionary
|
|
# make a copy
|
|
tmpobjs = objects.copy()
|
|
for k, v in objects.items():
|
|
parent_dc = self.get_parent_datacenter(k)
|
|
if parent_dc.name != self.dc_name:
|
|
tmpobjs.pop(k, None)
|
|
objects = tmpobjs
|
|
else:
|
|
# everything else should be a list
|
|
objects = [x for x in objects if self.get_parent_datacenter(x).name == self.dc_name]
|
|
|
|
return objects
|
|
|
|
def get_network(self, network):
|
|
if network not in self.networks:
|
|
self.networks[network] = self.find_obj(self.content, [vim.Network], network)
|
|
|
|
return self.networks[network]
|
|
|
|
def get_cluster(self, cluster):
|
|
if cluster not in self.clusters:
|
|
self.clusters[cluster] = self.find_obj(self.content, [vim.ClusterComputeResource], cluster)
|
|
|
|
return self.clusters[cluster]
|
|
|
|
def get_esx_host(self, host):
|
|
if host not in self.esx_hosts:
|
|
self.esx_hosts[host] = self.find_obj(self.content, [vim.HostSystem], host)
|
|
|
|
return self.esx_hosts[host]
|
|
|
|
def get_parent_datacenter(self, obj):
|
|
""" Walk the parent tree to find the objects datacenter """
|
|
if isinstance(obj, vim.Datacenter):
|
|
return obj
|
|
if obj in self.parent_datacenters:
|
|
return self.parent_datacenters[obj]
|
|
datacenter = None
|
|
while True:
|
|
if not hasattr(obj, 'parent'):
|
|
break
|
|
obj = obj.parent
|
|
if isinstance(obj, vim.Datacenter):
|
|
datacenter = obj
|
|
break
|
|
self.parent_datacenters[obj] = datacenter
|
|
return datacenter
|
|
|
|
|
|
class PyVmomiHelper(PyVmomi):
|
|
def __init__(self, module):
|
|
super(PyVmomiHelper, self).__init__(module)
|
|
self.device_helper = PyVmomiDeviceHelper(self.module)
|
|
self.configspec = None
|
|
self.change_detected = False
|
|
self.customspec = None
|
|
self.cache = PyVmomiCache(self.content, dc_name=self.params['datacenter'])
|
|
|
|
def gather_facts(self, vm):
|
|
return gather_vm_facts(self.content, vm)
|
|
|
|
def remove_vm(self, vm):
|
|
# https://www.vmware.com/support/developer/converter-sdk/conv60_apireference/vim.ManagedEntity.html#destroy
|
|
task = vm.Destroy()
|
|
self.wait_for_task(task)
|
|
|
|
if task.info.state == 'error':
|
|
return {'changed': False, 'failed': True, 'msg': task.info.error.msg}
|
|
else:
|
|
return {'changed': True, 'failed': False}
|
|
|
|
def configure_guestid(self, vm_obj, vm_creation=False):
|
|
# guest_id is not required when using templates
|
|
if self.params['template'] and not self.params['guest_id']:
|
|
return
|
|
|
|
# guest_id is only mandatory on VM creation
|
|
if vm_creation and self.params['guest_id'] is None:
|
|
self.module.fail_json(msg="guest_id attribute is mandatory for VM creation")
|
|
|
|
if self.params['guest_id'] and (vm_obj is None or self.params['guest_id'] != vm_obj.summary.config.guestId):
|
|
self.change_detected = True
|
|
self.configspec.guestId = self.params['guest_id']
|
|
|
|
def configure_cpu_and_memory(self, vm_obj, vm_creation=False):
|
|
# set cpu/memory/etc
|
|
if 'hardware' in self.params:
|
|
if 'num_cpus' in self.params['hardware']:
|
|
self.configspec.numCPUs = int(self.params['hardware']['num_cpus'])
|
|
if vm_obj is None or self.configspec.numCPUs != vm_obj.config.hardware.numCPU:
|
|
self.change_detected = True
|
|
# num_cpu is mandatory for VM creation
|
|
elif vm_creation and not self.params['template']:
|
|
self.module.fail_json(msg="hardware.num_cpus attribute is mandatory for VM creation")
|
|
|
|
if 'memory_mb' in self.params['hardware']:
|
|
self.configspec.memoryMB = int(self.params['hardware']['memory_mb'])
|
|
if vm_obj is None or self.configspec.memoryMB != vm_obj.config.hardware.memoryMB:
|
|
self.change_detected = True
|
|
# memory_mb is mandatory for VM creation
|
|
elif vm_creation and not self.params['template']:
|
|
self.module.fail_json(msg="hardware.memory_mb attribute is mandatory for VM creation")
|
|
|
|
def configure_cdrom(self, vm_obj):
|
|
# Configure the VM CD-ROM
|
|
if "cdrom" in self.params and self.params["cdrom"]:
|
|
if "type" not in self.params["cdrom"] or self.params["cdrom"]["type"] not in ["none", "client", "iso"]:
|
|
self.module.fail_json(msg="cdrom.type is mandatory")
|
|
if self.params["cdrom"]["type"] == "iso" and ("iso_path" not in self.params["cdrom"] or not self.params["cdrom"]["iso_path"]):
|
|
self.module.fail_json(msg="cdrom.iso_path is mandatory in case cdrom.type is iso")
|
|
|
|
if vm_obj and vm_obj.config.template:
|
|
# Changing CD-ROM settings on a template is not supported
|
|
return
|
|
|
|
cdrom_spec = None
|
|
cdrom_device = self.get_vm_cdrom_device(vm=vm_obj)
|
|
iso_path = self.params["cdrom"]["iso_path"] if "iso_path" in self.params["cdrom"] else None
|
|
if cdrom_device is None:
|
|
# Creating new CD-ROM
|
|
ide_device = self.get_vm_ide_device(vm=vm_obj)
|
|
if ide_device is None:
|
|
# Creating new IDE device
|
|
ide_device = self.device_helper.create_ide_controller()
|
|
self.change_detected = True
|
|
self.configspec.deviceChange.append(ide_device)
|
|
elif len(ide_device.device) > 3:
|
|
self.module.fail_json(msg="hardware.cdrom specified for a VM or template which already has 4 IDE devices of which none are a cdrom")
|
|
|
|
cdrom_spec = self.device_helper.create_cdrom(ide_ctl=ide_device, cdrom_type=self.params["cdrom"]["type"], iso_path=iso_path)
|
|
if vm_obj and vm_obj.runtime.powerState == vim.VirtualMachinePowerState.poweredOn:
|
|
cdrom_spec.device.connectable.connected = (self.params["cdrom"]["type"] != "none")
|
|
|
|
elif not self.device_helper.is_equal_cdrom(vm_obj=vm_obj, cdrom_device=cdrom_device, cdrom_type=self.params["cdrom"]["type"], iso_path=iso_path):
|
|
# Updating an existing CD-ROM
|
|
if self.params["cdrom"]["type"] in ["client", "none"]:
|
|
cdrom_device.backing = vim.vm.device.VirtualCdrom.RemotePassthroughBackingInfo()
|
|
elif self.params["cdrom"]["type"] == "iso":
|
|
cdrom_device.backing = vim.vm.device.VirtualCdrom.IsoBackingInfo(fileName=iso_path)
|
|
cdrom_device.connectable = vim.vm.device.VirtualDevice.ConnectInfo()
|
|
cdrom_device.connectable.allowGuestControl = True
|
|
cdrom_device.connectable.startConnected = (self.params["cdrom"]["type"] != "none")
|
|
if vm_obj and vm_obj.runtime.powerState == vim.VirtualMachinePowerState.poweredOn:
|
|
cdrom_device.connectable.connected = (self.params["cdrom"]["type"] != "none")
|
|
|
|
cdrom_spec = vim.vm.device.VirtualDeviceSpec()
|
|
cdrom_spec.operation = vim.vm.device.VirtualDeviceSpec.Operation.edit
|
|
cdrom_spec.device = cdrom_device
|
|
|
|
if cdrom_spec:
|
|
self.change_detected = True
|
|
self.configspec.deviceChange.append(cdrom_spec)
|
|
|
|
def get_vm_cdrom_device(self, vm=None):
|
|
if vm is None:
|
|
return None
|
|
|
|
for device in vm.config.hardware.device:
|
|
if isinstance(device, vim.vm.device.VirtualCdrom):
|
|
return device
|
|
|
|
return None
|
|
|
|
def get_vm_ide_device(self, vm=None):
|
|
if vm is None:
|
|
return None
|
|
|
|
for device in vm.config.hardware.device:
|
|
if isinstance(device, vim.vm.device.VirtualIDEController):
|
|
return device
|
|
|
|
return None
|
|
|
|
def get_vm_network_interfaces(self, vm=None):
|
|
if vm is None:
|
|
return []
|
|
|
|
device_list = []
|
|
for device in vm.config.hardware.device:
|
|
if isinstance(device, vim.vm.device.VirtualPCNet32) or \
|
|
isinstance(device, vim.vm.device.VirtualVmxnet2) or \
|
|
isinstance(device, vim.vm.device.VirtualVmxnet3) or \
|
|
isinstance(device, vim.vm.device.VirtualE1000) or \
|
|
isinstance(device, vim.vm.device.VirtualE1000e) or \
|
|
isinstance(device, vim.vm.device.VirtualSriovEthernetCard):
|
|
device_list.append(device)
|
|
|
|
return device_list
|
|
|
|
def configure_network(self, vm_obj):
|
|
# Ignore empty networks, this permits to keep networks when deploying a template/cloning a VM
|
|
if len(self.params['networks']) == 0:
|
|
return
|
|
|
|
network_devices = list()
|
|
for network in self.params['networks']:
|
|
if 'ip' in network or 'netmask' in network:
|
|
if 'ip' not in network or 'netmask' not in network:
|
|
self.module.fail_json(msg="Both 'ip' and 'netmask' are required together.")
|
|
|
|
if 'name' in network:
|
|
if find_obj(self.content, [vim.Network], network['name']) is None:
|
|
self.module.fail_json(msg="Network '%(name)s' does not exists" % network)
|
|
|
|
elif 'vlan' in network:
|
|
dvps = self.cache.get_all_objs(self.content, [vim.dvs.DistributedVirtualPortgroup])
|
|
for dvp in dvps:
|
|
if hasattr(dvp.config.defaultPortConfig, 'vlan') and dvp.config.defaultPortConfig.vlan.vlanId == network['vlan']:
|
|
network['name'] = dvp.config.name
|
|
break
|
|
if dvp.config.name == network['vlan']:
|
|
network['name'] = dvp.config.name
|
|
break
|
|
else:
|
|
self.module.fail_json(msg="VLAN '%(vlan)s' does not exist" % network)
|
|
else:
|
|
self.module.fail_json(msg="You need to define a network name or a vlan")
|
|
|
|
network_devices.append(network)
|
|
|
|
# List current device for Clone or Idempotency
|
|
current_net_devices = self.get_vm_network_interfaces(vm=vm_obj)
|
|
if len(network_devices) < len(current_net_devices):
|
|
self.module.fail_json(msg="given network device list is lesser than current VM device list (%d < %d). "
|
|
"Removing interfaces is not allowed"
|
|
% (len(network_devices), len(current_net_devices)))
|
|
|
|
for key in range(0, len(network_devices)):
|
|
# Default device type is vmxnet3, VMWare best practice
|
|
device_type = network_devices[key].get('device_type', 'vmxnet3')
|
|
nic = self.device_helper.create_nic(device_type,
|
|
'Network Adapter %s' % (key + 1),
|
|
network_devices[key])
|
|
|
|
nic_change_detected = False
|
|
if key < len(current_net_devices) and (vm_obj or self.params['template']):
|
|
nic.operation = vim.vm.device.VirtualDeviceSpec.Operation.edit
|
|
# Changing mac address has no effect when editing interface
|
|
if 'mac' in network_devices[key] and nic.device.macAddress != current_net_devices[key].macAddress:
|
|
self.module.fail_json(msg="Changing MAC address has not effect when interface is already present. "
|
|
"The failing new MAC address is %s" % nic.device.macAddress)
|
|
|
|
nic.device = current_net_devices[key]
|
|
nic.device.deviceInfo = vim.Description()
|
|
else:
|
|
nic.operation = vim.vm.device.VirtualDeviceSpec.Operation.add
|
|
nic_change_detected = True
|
|
|
|
if hasattr(self.cache.get_network(network_devices[key]['name']), 'portKeys'):
|
|
# VDS switch
|
|
pg_obj = find_obj(self.content, [vim.dvs.DistributedVirtualPortgroup], network_devices[key]['name'])
|
|
|
|
if (nic.device.backing and
|
|
(nic.device.backing.port.portgroupKey != pg_obj.key or
|
|
nic.device.backing.port.switchUuid != pg_obj.config.distributedVirtualSwitch.uuid)):
|
|
nic_change_detected = True
|
|
|
|
dvs_port_connection = vim.dvs.PortConnection()
|
|
dvs_port_connection.portgroupKey = pg_obj.key
|
|
dvs_port_connection.switchUuid = pg_obj.config.distributedVirtualSwitch.uuid
|
|
nic.device.backing = vim.vm.device.VirtualEthernetCard.DistributedVirtualPortBackingInfo()
|
|
nic.device.backing.port = dvs_port_connection
|
|
nic_change_detected = True
|
|
else:
|
|
# vSwitch
|
|
if not isinstance(nic.device.backing, vim.vm.device.VirtualEthernetCard.NetworkBackingInfo):
|
|
nic.device.backing = vim.vm.device.VirtualEthernetCard.NetworkBackingInfo()
|
|
nic_change_detected = True
|
|
|
|
net_obj = self.cache.get_network(network_devices[key]['name'])
|
|
if nic.device.backing.network != net_obj:
|
|
nic.device.backing.network = net_obj
|
|
nic_change_detected = True
|
|
|
|
if nic.device.backing.deviceName != network_devices[key]['name']:
|
|
nic.device.backing.deviceName = network_devices[key]['name']
|
|
nic_change_detected = True
|
|
|
|
if nic_change_detected:
|
|
self.configspec.deviceChange.append(nic)
|
|
self.change_detected = True
|
|
|
|
def customize_customvalues(self, vm_obj):
|
|
if len(self.params['customvalues']) == 0:
|
|
return
|
|
|
|
facts = self.gather_facts(vm_obj)
|
|
for kv in self.params['customvalues']:
|
|
if 'key' not in kv or 'value' not in kv:
|
|
self.module.exit_json(msg="customvalues items required both 'key' and 'value fields.")
|
|
|
|
# If kv is not kv fetched from facts, change it
|
|
if kv['key'] not in facts['customvalues'] or facts['customvalues'][kv['key']] != kv['value']:
|
|
try:
|
|
vm_obj.setCustomValue(key=kv['key'], value=kv['value'])
|
|
self.change_detected = True
|
|
except Exception as e:
|
|
self.module.fail_json(msg="Failed to set custom value for key='%s' and value='%s'. Error was: %s"
|
|
% (kv['key'], kv['value'], to_text(e)))
|
|
|
|
def customize_vm(self, vm_obj):
|
|
# Network settings
|
|
adaptermaps = []
|
|
for network in self.params['networks']:
|
|
|
|
guest_map = vim.vm.customization.AdapterMapping()
|
|
guest_map.adapter = vim.vm.customization.IPSettings()
|
|
|
|
if 'ip' in network and 'netmask' in network:
|
|
if 'type' in network and network['type'] != 'static':
|
|
self.module.fail_json(msg='Static IP information provided for network "%(name)s", but "type" is set to "%(type)s".' % network)
|
|
guest_map.adapter.ip = vim.vm.customization.FixedIp()
|
|
guest_map.adapter.ip.ipAddress = str(network['ip'])
|
|
guest_map.adapter.subnetMask = str(network['netmask'])
|
|
elif 'type' in network and network['type'] == 'static':
|
|
self.module.fail_json(msg='Network "%(name)s" was set to type "%(type)s", but "ip" and "netmask" are missing.' % network)
|
|
elif 'type' in network and network['type'] == 'dhcp':
|
|
guest_map.adapter.ip = vim.vm.customization.DhcpIpGenerator()
|
|
else:
|
|
self.module.fail_json(msg='Network "%(name)s" was set to unknown type "%(type)s".' % network)
|
|
|
|
if 'gateway' in network:
|
|
guest_map.adapter.gateway = network['gateway']
|
|
|
|
# On Windows, DNS domain and DNS servers can be set by network interface
|
|
# https://pubs.vmware.com/vi3/sdk/ReferenceGuide/vim.vm.customization.IPSettings.html
|
|
if 'domain' in network:
|
|
guest_map.adapter.dnsDomain = network['domain']
|
|
elif 'domain' in self.params['customization']:
|
|
guest_map.adapter.dnsDomain = self.params['customization']['domain']
|
|
|
|
if 'dns_servers' in network:
|
|
guest_map.adapter.dnsServerList = network['dns_servers']
|
|
elif 'dns_servers' in self.params['customization']:
|
|
guest_map.adapter.dnsServerList = self.params['customization']['dns_servers']
|
|
|
|
adaptermaps.append(guest_map)
|
|
|
|
# Global DNS settings
|
|
globalip = vim.vm.customization.GlobalIPSettings()
|
|
if 'dns_servers' in self.params['customization']:
|
|
globalip.dnsServerList = self.params['customization']['dns_servers']
|
|
|
|
# TODO: Maybe list the different domains from the interfaces here by default ?
|
|
if 'dns_suffix' in self.params['customization']:
|
|
globalip.dnsSuffixList = self.params['customization']['dns_suffix']
|
|
elif 'domain' in self.params['customization']:
|
|
globalip.dnsSuffixList = self.params['customization']['domain']
|
|
|
|
if self.params['guest_id']:
|
|
guest_id = self.params['guest_id']
|
|
else:
|
|
guest_id = vm_obj.summary.config.guestId
|
|
|
|
# For windows guest OS, use SysPrep
|
|
# https://pubs.vmware.com/vi3/sdk/ReferenceGuide/vim.vm.customization.Sysprep.html#field_detail
|
|
if 'win' in guest_id:
|
|
ident = vim.vm.customization.Sysprep()
|
|
|
|
ident.userData = vim.vm.customization.UserData()
|
|
|
|
# Setting hostName, orgName and fullName is mandatory, so we set some default when missing
|
|
ident.userData.computerName = vim.vm.customization.FixedName()
|
|
ident.userData.computerName.name = str(self.params['customization'].get('hostname', self.params['name'].split('.')[0]))
|
|
ident.userData.fullName = str(self.params['customization'].get('fullname', 'Administrator'))
|
|
ident.userData.orgName = str(self.params['customization'].get('orgname', 'ACME'))
|
|
|
|
if 'productid' in self.params['customization']:
|
|
ident.userData.productId = str(self.params['customization']['productid'])
|
|
|
|
ident.guiUnattended = vim.vm.customization.GuiUnattended()
|
|
|
|
if 'autologon' in self.params['customization']:
|
|
ident.guiUnattended.autoLogon = self.params['customization']['autologon']
|
|
ident.guiUnattended.autoLogonCount = self.params['customization'].get('autologoncount', 1)
|
|
|
|
if 'timezone' in self.params['customization']:
|
|
ident.guiUnattended.timeZone = self.params['customization']['timezone']
|
|
|
|
ident.identification = vim.vm.customization.Identification()
|
|
|
|
if self.params['customization'].get('password', '') != '':
|
|
ident.guiUnattended.password = vim.vm.customization.Password()
|
|
ident.guiUnattended.password.value = str(self.params['customization']['password'])
|
|
ident.guiUnattended.password.plainText = True
|
|
|
|
if 'joindomain' in self.params['customization']:
|
|
if 'domainadmin' not in self.params['customization'] or 'domainadminpassword' not in self.params['customization']:
|
|
self.module.fail_json(msg="'domainadmin' and 'domainadminpassword' entries are mandatory in 'customization' section to use "
|
|
"joindomain feature")
|
|
|
|
ident.identification.domainAdmin = str(self.params['customization']['domainadmin'])
|
|
ident.identification.joinDomain = str(self.params['customization']['joindomain'])
|
|
ident.identification.domainAdminPassword = vim.vm.customization.Password()
|
|
ident.identification.domainAdminPassword.value = str(self.params['customization']['domainadminpassword'])
|
|
ident.identification.domainAdminPassword.plainText = True
|
|
|
|
elif 'joinworkgroup' in self.params['customization']:
|
|
ident.identification.joinWorkgroup = str(self.params['customization']['joinworkgroup'])
|
|
|
|
if 'runonce' in self.params['customization']:
|
|
ident.guiRunOnce = vim.vm.customization.GuiRunOnce()
|
|
ident.guiRunOnce.commandList = self.params['customization']['runonce']
|
|
|
|
else:
|
|
# FIXME: We have no clue whether this non-Windows OS is actually Linux, hence it might fail !
|
|
|
|
# For Linux guest OS, use LinuxPrep
|
|
# https://pubs.vmware.com/vi3/sdk/ReferenceGuide/vim.vm.customization.LinuxPrep.html
|
|
ident = vim.vm.customization.LinuxPrep()
|
|
|
|
# TODO: Maybe add domain from interface if missing ?
|
|
if 'domain' in self.params['customization']:
|
|
ident.domain = str(self.params['customization']['domain'])
|
|
|
|
ident.hostName = vim.vm.customization.FixedName()
|
|
ident.hostName.name = str(self.params['customization'].get('hostname', self.params['name'].split('.')[0]))
|
|
|
|
self.customspec = vim.vm.customization.Specification()
|
|
self.customspec.nicSettingMap = adaptermaps
|
|
self.customspec.globalIPSettings = globalip
|
|
self.customspec.identity = ident
|
|
|
|
def get_vm_scsi_controller(self, vm_obj):
|
|
# If vm_obj doesn't exists no SCSI controller to find
|
|
if vm_obj is None:
|
|
return None
|
|
|
|
for device in vm_obj.config.hardware.device:
|
|
if self.device_helper.is_scsi_controller(device):
|
|
scsi_ctl = vim.vm.device.VirtualDeviceSpec()
|
|
scsi_ctl.device = device
|
|
return scsi_ctl
|
|
|
|
return None
|
|
|
|
def get_configured_disk_size(self, expected_disk_spec):
|
|
# what size is it?
|
|
if [x for x in expected_disk_spec.keys() if x.startswith('size_') or x == 'size']:
|
|
# size_tb, size_gb, size_mb, size_kb, size_b ...?
|
|
if 'size' in expected_disk_spec:
|
|
expected = ''.join(c for c in expected_disk_spec['size'] if c.isdigit())
|
|
unit = expected_disk_spec['size'].replace(expected, '').lower()
|
|
expected = int(expected)
|
|
else:
|
|
param = [x for x in expected_disk_spec.keys() if x.startswith('size_')][0]
|
|
unit = param.split('_')[-1].lower()
|
|
expected = [x[1] for x in expected_disk_spec.items() if x[0].startswith('size_')][0]
|
|
expected = int(expected)
|
|
|
|
if unit == 'tb':
|
|
return expected * 1024 * 1024 * 1024
|
|
elif unit == 'gb':
|
|
return expected * 1024 * 1024
|
|
elif unit == 'mb':
|
|
return expected * 1024
|
|
elif unit == 'kb':
|
|
return expected
|
|
|
|
self.module.fail_json(
|
|
msg='%s is not a supported unit for disk size. Supported units are kb, mb, gb or tb' % unit)
|
|
|
|
# No size found but disk, fail
|
|
self.module.fail_json(
|
|
msg="No size, size_kb, size_mb, size_gb or size_tb attribute found into disk configuration")
|
|
|
|
def configure_disks(self, vm_obj):
|
|
# Ignore empty disk list, this permits to keep disks when deploying a template/cloning a VM
|
|
if len(self.params['disk']) == 0:
|
|
return
|
|
|
|
scsi_ctl = self.get_vm_scsi_controller(vm_obj)
|
|
|
|
# Create scsi controller only if we are deploying a new VM, not a template or reconfiguring
|
|
if vm_obj is None or scsi_ctl is None:
|
|
scsi_ctl = self.device_helper.create_scsi_controller(self.get_scsi_type())
|
|
self.change_detected = True
|
|
self.configspec.deviceChange.append(scsi_ctl)
|
|
|
|
disks = [x for x in vm_obj.config.hardware.device if isinstance(x, vim.vm.device.VirtualDisk)] \
|
|
if vm_obj is not None else None
|
|
|
|
if disks is not None and self.params.get('disk') and len(self.params.get('disk')) < len(disks):
|
|
self.module.fail_json(msg="Provided disks configuration has less disks than "
|
|
"the target object (%d vs %d)" % (len(self.params.get('disk')), len(disks)))
|
|
|
|
disk_index = 0
|
|
for expected_disk_spec in self.params.get('disk'):
|
|
disk_modified = False
|
|
# If we are manipulating and existing objects which has disks and disk_index is in disks
|
|
if vm_obj is not None and disks is not None and disk_index < len(disks):
|
|
diskspec = vim.vm.device.VirtualDeviceSpec()
|
|
# set the operation to edit so that it knows to keep other settings
|
|
diskspec.operation = vim.vm.device.VirtualDeviceSpec.Operation.edit
|
|
diskspec.device = disks[disk_index]
|
|
else:
|
|
diskspec = self.device_helper.create_scsi_disk(scsi_ctl, disk_index)
|
|
disk_modified = True
|
|
|
|
# is it thin?
|
|
if 'type' in expected_disk_spec:
|
|
if expected_disk_spec.get('type', '').lower() == 'thin':
|
|
diskspec.device.backing.thinProvisioned = True
|
|
|
|
# which datastore?
|
|
if expected_disk_spec.get('datastore'):
|
|
# TODO: This is already handled by the relocation spec,
|
|
# but it needs to eventually be handled for all the
|
|
# other disks defined
|
|
pass
|
|
|
|
# increment index for next disk search
|
|
disk_index += 1
|
|
# index 7 is reserved to SCSI controller
|
|
if disk_index == 7:
|
|
disk_index += 1
|
|
|
|
kb = self.get_configured_disk_size(expected_disk_spec)
|
|
# VMWare doesn't allow to reduce disk sizes
|
|
if kb < diskspec.device.capacityInKB:
|
|
self.module.fail_json(
|
|
msg="Given disk size is smaller than found (%d < %d). Reducing disks is not allowed." %
|
|
(kb, diskspec.device.capacityInKB))
|
|
|
|
if kb != diskspec.device.capacityInKB or disk_modified:
|
|
diskspec.device.capacityInKB = kb
|
|
self.configspec.deviceChange.append(diskspec)
|
|
|
|
self.change_detected = True
|
|
|
|
def select_host(self):
|
|
# if the user wants a cluster, get the list of hosts for the cluster and use the first one
|
|
if self.params['cluster']:
|
|
cluster = self.cache.get_cluster(self.params['cluster'])
|
|
if not cluster:
|
|
self.module.fail_json(msg='Failed to find cluster "%(cluster)s"' % self.params)
|
|
hostsystems = [x for x in cluster.host]
|
|
if not hostsystems:
|
|
self.module.fail_json(msg='No hosts found in cluster "%(cluster)s. Maybe you lack the right privileges ?"' % self.params)
|
|
# TODO: add a policy to select host
|
|
hostsystem = hostsystems[0]
|
|
else:
|
|
hostsystem = self.cache.get_esx_host(self.params['esxi_hostname'])
|
|
if not hostsystem:
|
|
self.module.fail_json(msg='Failed to find ESX host "%(esxi_hostname)s"' % self.params)
|
|
|
|
return hostsystem
|
|
|
|
def autoselect_datastore(self):
|
|
datastore = None
|
|
datastores = self.cache.get_all_objs(self.content, [vim.Datastore])
|
|
|
|
if datastores is None or len(datastores) == 0:
|
|
self.module.fail_json(msg="Unable to find a datastore list when autoselecting")
|
|
|
|
datastore_freespace = 0
|
|
for ds in datastores:
|
|
if ds.summary.freeSpace > datastore_freespace:
|
|
datastore = ds
|
|
datastore_freespace = ds.summary.freeSpace
|
|
|
|
return datastore
|
|
|
|
def select_datastore(self, vm_obj=None):
|
|
datastore = None
|
|
datastore_name = None
|
|
|
|
if len(self.params['disk']) != 0:
|
|
# TODO: really use the datastore for newly created disks
|
|
if 'autoselect_datastore' in self.params['disk'][0] and self.params['disk'][0]['autoselect_datastore']:
|
|
datastores = self.cache.get_all_objs(self.content, [vim.Datastore])
|
|
datastores = [x for x in datastores if self.cache.get_parent_datacenter(x).name == self.params['datacenter']]
|
|
if datastores is None or len(datastores) == 0:
|
|
self.module.fail_json(msg="Unable to find a datastore list when autoselecting")
|
|
|
|
datastore_freespace = 0
|
|
for ds in datastores:
|
|
if (ds.summary.freeSpace > datastore_freespace) or (ds.summary.freeSpace == datastore_freespace and not datastore):
|
|
# If datastore field is provided, filter destination datastores
|
|
if 'datastore' in self.params['disk'][0] and \
|
|
isinstance(self.params['disk'][0]['datastore'], str) and \
|
|
ds.name.find(self.params['disk'][0]['datastore']) < 0:
|
|
continue
|
|
|
|
datastore = ds
|
|
datastore_name = datastore.name
|
|
datastore_freespace = ds.summary.freeSpace
|
|
|
|
elif 'datastore' in self.params['disk'][0]:
|
|
datastore_name = self.params['disk'][0]['datastore']
|
|
datastore = self.cache.find_obj(self.content, [vim.Datastore], datastore_name)
|
|
else:
|
|
self.module.fail_json(msg="Either datastore or autoselect_datastore should be provided to select datastore")
|
|
|
|
if not datastore and self.params['template']:
|
|
# use the template's existing DS
|
|
disks = [x for x in vm_obj.config.hardware.device if isinstance(x, vim.vm.device.VirtualDisk)]
|
|
if disks:
|
|
datastore = disks[0].backing.datastore
|
|
datastore_name = datastore.name
|
|
# validation
|
|
if datastore:
|
|
dc = self.cache.get_parent_datacenter(datastore)
|
|
if dc.name != self.params['datacenter']:
|
|
datastore = self.autoselect_datastore()
|
|
datastore_name = datastore.name
|
|
|
|
if not datastore:
|
|
self.module.fail_json(msg="Failed to find a matching datastore")
|
|
|
|
return datastore, datastore_name
|
|
|
|
def obj_has_parent(self, obj, parent):
|
|
assert obj is not None and parent is not None
|
|
current_parent = obj
|
|
|
|
while True:
|
|
if current_parent.name == parent.name:
|
|
return True
|
|
|
|
current_parent = current_parent.parent
|
|
if current_parent is None:
|
|
return False
|
|
|
|
def select_resource_pool_by_name(self, resource_pool_name):
|
|
resource_pool = self.cache.find_obj(self.content, [vim.ResourcePool], resource_pool_name)
|
|
if resource_pool is None:
|
|
self.module.fail_json(msg='Could not find resource_pool "%s"' % resource_pool_name)
|
|
return resource_pool
|
|
|
|
def select_resource_pool_by_host(self, host):
|
|
resource_pools = self.cache.get_all_objs(self.content, [vim.ResourcePool])
|
|
for rp in resource_pools.items():
|
|
if not rp[0]:
|
|
continue
|
|
|
|
if not hasattr(rp[0], 'parent') or not rp[0].parent:
|
|
continue
|
|
|
|
# Find resource pool on host
|
|
if self.obj_has_parent(rp[0].parent, host.parent):
|
|
# If no resource_pool selected or it's the selected pool, return it
|
|
if self.module.params['resource_pool'] is None or rp[0].name == self.module.params['resource_pool']:
|
|
return rp[0]
|
|
|
|
if self.module.params['resource_pool'] is not None:
|
|
self.module.fail_json(msg="Could not find resource_pool %s for selected host %s"
|
|
% (self.module.params['resource_pool'], host.name))
|
|
else:
|
|
self.module.fail_json(msg="Failed to find a resource group for %s" % host.name)
|
|
|
|
def get_scsi_type(self):
|
|
disk_controller_type = "paravirtual"
|
|
# set cpu/memory/etc
|
|
if 'hardware' in self.params:
|
|
if 'scsi' in self.params['hardware']:
|
|
if self.params['hardware']['scsi'] in ['buslogic', 'paravirtual', 'lsilogic', 'lsilogicsas']:
|
|
disk_controller_type = self.params['hardware']['scsi']
|
|
else:
|
|
self.module.fail_json(msg="hardware.scsi attribute should be 'paravirtual' or 'lsilogic'")
|
|
return disk_controller_type
|
|
|
|
def find_folder(self, searchpath):
|
|
""" Walk inventory objects one position of the searchpath at a time """
|
|
|
|
# split the searchpath so we can iterate through it
|
|
paths = [x.replace('/', '') for x in searchpath.split('/')]
|
|
paths_total = len(paths) - 1
|
|
position = 0
|
|
|
|
# recursive walk while looking for next element in searchpath
|
|
root = self.content.rootFolder
|
|
while root and position <= paths_total:
|
|
change = False
|
|
if hasattr(root, 'childEntity'):
|
|
for child in root.childEntity:
|
|
if child.name == paths[position]:
|
|
root = child
|
|
position += 1
|
|
change = True
|
|
break
|
|
elif isinstance(root, vim.Datacenter):
|
|
if hasattr(root, 'vmFolder'):
|
|
if root.vmFolder.name == paths[position]:
|
|
root = root.vmFolder
|
|
position += 1
|
|
change = True
|
|
else:
|
|
root = None
|
|
|
|
if not change:
|
|
root = None
|
|
|
|
return root
|
|
|
|
def get_resource_pool(self):
|
|
|
|
resource_pool = None
|
|
|
|
if self.params['esxi_hostname']:
|
|
host = self.select_host()
|
|
resource_pool = self.select_resource_pool_by_host(host)
|
|
else:
|
|
resource_pool = self.select_resource_pool_by_name(self.params['resource_pool'])
|
|
|
|
if resource_pool is None:
|
|
self.module.fail_json(msg='Unable to find resource pool "%(resource_pool)s"' % self.params)
|
|
|
|
return resource_pool
|
|
|
|
def deploy_vm(self):
|
|
# https://github.com/vmware/pyvmomi-community-samples/blob/master/samples/clone_vm.py
|
|
# https://www.vmware.com/support/developer/vc-sdk/visdk25pubs/ReferenceGuide/vim.vm.CloneSpec.html
|
|
# https://www.vmware.com/support/developer/vc-sdk/visdk25pubs/ReferenceGuide/vim.vm.ConfigSpec.html
|
|
# https://www.vmware.com/support/developer/vc-sdk/visdk41pubs/ApiReference/vim.vm.RelocateSpec.html
|
|
|
|
# FIXME:
|
|
# - multiple templates by the same name
|
|
# - static IPs
|
|
|
|
# datacenters = get_all_objs(self.content, [vim.Datacenter])
|
|
datacenter = self.cache.find_obj(self.content, [vim.Datacenter], self.params['datacenter'])
|
|
if datacenter is None:
|
|
self.module.fail_json(msg='No datacenter named %(datacenter)s was found' % self.params)
|
|
|
|
# Prepend / if it was missing from the folder path, also strip trailing slashes
|
|
if not self.params['folder'].startswith('/'):
|
|
self.params['folder'] = '/%(folder)s' % self.params
|
|
self.params['folder'] = self.params['folder'].rstrip('/')
|
|
|
|
dcpath = compile_folder_path_for_object(datacenter)
|
|
|
|
# Check for full path first in case it was already supplied
|
|
if (self.params['folder'].startswith(dcpath + self.params['datacenter'] + '/vm') or
|
|
self.params['folder'].startswith(dcpath + '/' + self.params['datacenter'] + '/vm')):
|
|
fullpath = self.params['folder']
|
|
elif (self.params['folder'].startswith('/vm/') or self.params['folder'] == '/vm'):
|
|
fullpath = "%s%s%s" % (dcpath, self.params['datacenter'], self.params['folder'])
|
|
elif (self.params['folder'].startswith('/')):
|
|
fullpath = "%s%s/vm%s" % (dcpath, self.params['datacenter'], self.params['folder'])
|
|
else:
|
|
fullpath = "%s%s/vm/%s" % (dcpath, self.params['datacenter'], self.params['folder'])
|
|
|
|
f_obj = self.content.searchIndex.FindByInventoryPath(fullpath)
|
|
|
|
# abort if no strategy was successful
|
|
if f_obj is None:
|
|
self.module.fail_json(msg='No folder matched the path: %(folder)s' % self.params)
|
|
destfolder = f_obj
|
|
|
|
if self.params['template']:
|
|
# FIXME: need to search for this in the same way as guests to ensure accuracy
|
|
vm_obj = find_obj(self.content, [vim.VirtualMachine], self.params['template'])
|
|
if vm_obj is None:
|
|
self.module.fail_json(msg="Could not find a template named %(template)s" % self.params)
|
|
else:
|
|
vm_obj = None
|
|
|
|
# need a resource pool if cloning from template
|
|
if self.params['resource_pool'] or self.params['template']:
|
|
resource_pool = self.get_resource_pool()
|
|
|
|
# set the destination datastore for VM & disks
|
|
(datastore, datastore_name) = self.select_datastore(vm_obj)
|
|
|
|
self.configspec = vim.vm.ConfigSpec(cpuHotAddEnabled=True, memoryHotAddEnabled=True)
|
|
self.configspec.deviceChange = []
|
|
self.configure_guestid(vm_obj=vm_obj, vm_creation=True)
|
|
self.configure_cpu_and_memory(vm_obj=vm_obj, vm_creation=True)
|
|
self.configure_disks(vm_obj=vm_obj)
|
|
self.configure_network(vm_obj=vm_obj)
|
|
self.configure_cdrom(vm_obj=vm_obj)
|
|
|
|
# Find if we need network customizations (find keys in dictionary that requires customizations)
|
|
network_changes = False
|
|
for nw in self.params['networks']:
|
|
for key in nw:
|
|
# We don't need customizations for these keys
|
|
if key not in ('device_type', 'mac', 'name', 'vlan'):
|
|
network_changes = True
|
|
break
|
|
|
|
if len(self.params['customization']) > 0 or network_changes is True:
|
|
self.customize_vm(vm_obj=vm_obj)
|
|
|
|
clonespec = None
|
|
clone_method = None
|
|
try:
|
|
if self.params['template']:
|
|
# create the relocation spec
|
|
relospec = vim.vm.RelocateSpec()
|
|
|
|
# Only select specific host when ESXi hostname is provided
|
|
if self.params['esxi_hostname']:
|
|
relospec.host = self.select_host()
|
|
relospec.datastore = datastore
|
|
|
|
# https://www.vmware.com/support/developer/vc-sdk/visdk41pubs/ApiReference/vim.vm.RelocateSpec.html
|
|
# > pool: For a clone operation from a template to a virtual machine, this argument is required.
|
|
relospec.pool = resource_pool
|
|
|
|
if self.params['snapshot_src'] is not None and self.params['linked_clone']:
|
|
relospec.diskMoveType = vim.vm.RelocateSpec.DiskMoveOptions.createNewChildDiskBacking
|
|
|
|
clonespec = vim.vm.CloneSpec(template=self.params['is_template'], location=relospec)
|
|
if self.customspec:
|
|
clonespec.customization = self.customspec
|
|
|
|
if self.params['snapshot_src'] is not None:
|
|
snapshot = self.get_snapshots_by_name_recursively(snapshots=vm_obj.snapshot.rootSnapshotList, snapname=self.params['snapshot_src'])
|
|
if len(snapshot) != 1:
|
|
self.module.fail_json(msg='virtual machine "%(template)s" does not contain snapshot named "%(snapshot_src)s"' % self.params)
|
|
|
|
clonespec.snapshot = snapshot[0].snapshot
|
|
|
|
clonespec.config = self.configspec
|
|
clone_method = 'Clone'
|
|
task = vm_obj.Clone(folder=destfolder, name=self.params['name'], spec=clonespec)
|
|
self.change_detected = True
|
|
else:
|
|
# ConfigSpec require name for VM creation
|
|
self.configspec.name = self.params['name']
|
|
self.configspec.files = vim.vm.FileInfo(logDirectory=None,
|
|
snapshotDirectory=None,
|
|
suspendDirectory=None,
|
|
vmPathName="[" + datastore_name + "] " + self.params["name"])
|
|
|
|
clone_method = 'CreateVM_Task'
|
|
resource_pool = self.get_resource_pool()
|
|
task = destfolder.CreateVM_Task(config=self.configspec, pool=resource_pool)
|
|
self.change_detected = True
|
|
self.wait_for_task(task)
|
|
except TypeError as e:
|
|
self.module.fail_json(msg="TypeError was returned, please ensure to give correct inputs. %s" % to_text(e))
|
|
|
|
if task.info.state == 'error':
|
|
# https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2021361
|
|
# https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2173
|
|
|
|
# provide these to the user for debugging
|
|
clonespec_json = serialize_spec(clonespec)
|
|
configspec_json = serialize_spec(self.configspec)
|
|
kwargs = {
|
|
'changed': self.change_detected,
|
|
'failed': True,
|
|
'msg': task.info.error.msg,
|
|
'clonespec': clonespec_json,
|
|
'configspec': configspec_json,
|
|
'clone_method': clone_method
|
|
}
|
|
|
|
return kwargs
|
|
else:
|
|
# set annotation
|
|
vm = task.info.result
|
|
if self.params['annotation']:
|
|
annotation_spec = vim.vm.ConfigSpec()
|
|
annotation_spec.annotation = str(self.params['annotation'])
|
|
task = vm.ReconfigVM_Task(annotation_spec)
|
|
self.wait_for_task(task)
|
|
|
|
self.customize_customvalues(vm_obj=vm)
|
|
|
|
if self.params['wait_for_ip_address'] or self.params['state'] in ['poweredon', 'restarted']:
|
|
set_vm_power_state(self.content, vm, 'poweredon', force=False)
|
|
|
|
if self.params['wait_for_ip_address']:
|
|
self.wait_for_vm_ip(vm)
|
|
|
|
vm_facts = self.gather_facts(vm)
|
|
return {'changed': self.change_detected, 'failed': False, 'instance': vm_facts}
|
|
|
|
def get_snapshots_by_name_recursively(self, snapshots, snapname):
|
|
snap_obj = []
|
|
for snapshot in snapshots:
|
|
if snapshot.name == snapname:
|
|
snap_obj.append(snapshot)
|
|
else:
|
|
snap_obj = snap_obj + self.get_snapshots_by_name_recursively(snapshot.childSnapshotList, snapname)
|
|
return snap_obj
|
|
|
|
def reconfigure_vm(self):
|
|
self.configspec = vim.vm.ConfigSpec()
|
|
self.configspec.deviceChange = []
|
|
|
|
self.configure_guestid(vm_obj=self.current_vm_obj)
|
|
self.configure_cpu_and_memory(vm_obj=self.current_vm_obj)
|
|
self.configure_disks(vm_obj=self.current_vm_obj)
|
|
self.configure_network(vm_obj=self.current_vm_obj)
|
|
self.configure_cdrom(vm_obj=self.current_vm_obj)
|
|
self.customize_customvalues(vm_obj=self.current_vm_obj)
|
|
|
|
if self.params['annotation'] and self.current_vm_obj.config.annotation != self.params['annotation']:
|
|
self.configspec.annotation = str(self.params['annotation'])
|
|
self.change_detected = True
|
|
|
|
change_applied = False
|
|
|
|
relospec = vim.vm.RelocateSpec()
|
|
if self.params['resource_pool']:
|
|
relospec.pool = self.select_resource_pool_by_name(self.params['resource_pool'])
|
|
|
|
if relospec.pool is None:
|
|
self.module.fail_json(msg='Unable to find resource pool "%(resource_pool)s"' % self.params)
|
|
|
|
elif relospec.pool != self.current_vm_obj.resourcePool:
|
|
task = self.current_vm_obj.RelocateVM_Task(spec=relospec)
|
|
self.wait_for_task(task)
|
|
change_applied = True
|
|
|
|
# Only send VMWare task if we see a modification
|
|
if self.change_detected:
|
|
task = self.current_vm_obj.ReconfigVM_Task(spec=self.configspec)
|
|
self.wait_for_task(task)
|
|
change_applied = True
|
|
|
|
if task.info.state == 'error':
|
|
# https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2021361
|
|
# https://kb.vmware.com/selfservice/microsites/search.do?language=en_US&cmd=displayKC&externalId=2173
|
|
return {'changed': change_applied, 'failed': True, 'msg': task.info.error.msg}
|
|
|
|
# Rename VM
|
|
if self.params['uuid'] and self.params['name'] and self.params['name'] != self.current_vm_obj.config.name:
|
|
task = self.current_vm_obj.Rename_Task(self.params['name'])
|
|
self.wait_for_task(task)
|
|
change_applied = True
|
|
|
|
if task.info.state == 'error':
|
|
return {'changed': change_applied, 'failed': True, 'msg': task.info.error.msg}
|
|
|
|
# Mark VM as Template
|
|
if self.params['is_template'] and not self.current_vm_obj.config.template:
|
|
self.current_vm_obj.MarkAsTemplate()
|
|
change_applied = True
|
|
|
|
# Mark Template as VM
|
|
elif not self.params['is_template'] and self.current_vm_obj.config.template:
|
|
if self.params['resource_pool']:
|
|
resource_pool = self.select_resource_pool_by_name(self.params['resource_pool'])
|
|
|
|
if resource_pool is None:
|
|
self.module.fail_json(msg='Unable to find resource pool "%(resource_pool)s"' % self.params)
|
|
|
|
self.current_vm_obj.MarkAsVirtualMachine(pool=resource_pool)
|
|
|
|
# Automatically update VMWare UUID when converting template to VM.
|
|
# This avoids an interactive prompt during VM startup.
|
|
uuid_action = [x for x in self.current_vm_obj.config.extraConfig if x.key == "uuid.action"]
|
|
if not uuid_action:
|
|
uuid_action_opt = vim.option.OptionValue()
|
|
uuid_action_opt.key = "uuid.action"
|
|
uuid_action_opt.value = "create"
|
|
self.configspec.extraConfig.append(uuid_action_opt)
|
|
self.change_detected = True
|
|
|
|
change_applied = True
|
|
else:
|
|
self.module.fail_json(msg="Resource pool must be specified when converting template to VM!")
|
|
|
|
vm_facts = self.gather_facts(self.current_vm_obj)
|
|
return {'changed': change_applied, 'failed': False, 'instance': vm_facts}
|
|
|
|
@staticmethod
|
|
def wait_for_task(task):
|
|
# https://www.vmware.com/support/developer/vc-sdk/visdk25pubs/ReferenceGuide/vim.Task.html
|
|
# https://www.vmware.com/support/developer/vc-sdk/visdk25pubs/ReferenceGuide/vim.TaskInfo.html
|
|
# https://github.com/virtdevninja/pyvmomi-community-samples/blob/master/samples/tools/tasks.py
|
|
while task.info.state not in ['error', 'success']:
|
|
time.sleep(1)
|
|
|
|
def wait_for_vm_ip(self, vm, poll=100, sleep=5):
|
|
ips = None
|
|
facts = {}
|
|
thispoll = 0
|
|
while not ips and thispoll <= poll:
|
|
newvm = self.get_vm()
|
|
facts = self.gather_facts(newvm)
|
|
if facts['ipv4'] or facts['ipv6']:
|
|
ips = True
|
|
else:
|
|
time.sleep(sleep)
|
|
thispoll += 1
|
|
|
|
return facts
|
|
|
|
|
|
def main():
|
|
argument_spec = vmware_argument_spec()
|
|
argument_spec.update(
|
|
state=dict(type='str', default='present',
|
|
choices=['absent', 'poweredoff', 'poweredon', 'present', 'rebootguest', 'restarted', 'shutdownguest', 'suspended']),
|
|
template=dict(type='str', aliases=['template_src']),
|
|
is_template=dict(type='bool', default=False),
|
|
annotation=dict(type='str', aliases=['notes']),
|
|
customvalues=dict(type='list', default=[]),
|
|
name=dict(type='str', required=True),
|
|
name_match=dict(type='str', choices=['first', 'last'], default='first'),
|
|
uuid=dict(type='str'),
|
|
folder=dict(type='str', default='/vm'),
|
|
guest_id=dict(type='str'),
|
|
disk=dict(type='list', default=[]),
|
|
cdrom=dict(type='dict', default={}),
|
|
hardware=dict(type='dict', default={}),
|
|
force=dict(type='bool', default=False),
|
|
datacenter=dict(type='str', default='ha-datacenter'),
|
|
esxi_hostname=dict(type='str'),
|
|
cluster=dict(type='str'),
|
|
wait_for_ip_address=dict(type='bool', default=False),
|
|
snapshot_src=dict(type='str'),
|
|
linked_clone=dict(type='bool', default=False),
|
|
networks=dict(type='list', default=[]),
|
|
resource_pool=dict(type='str'),
|
|
customization=dict(type='dict', default={}, no_log=True),
|
|
)
|
|
|
|
module = AnsibleModule(argument_spec=argument_spec,
|
|
supports_check_mode=True,
|
|
mutually_exclusive=[
|
|
['cluster', 'esxi_hostname'],
|
|
],
|
|
)
|
|
|
|
result = {'failed': False, 'changed': False}
|
|
|
|
# FindByInventoryPath() does not require an absolute path
|
|
# so we should leave the input folder path unmodified
|
|
module.params['folder'] = module.params['folder'].rstrip('/')
|
|
|
|
pyv = PyVmomiHelper(module)
|
|
|
|
# Check if the VM exists before continuing
|
|
vm = pyv.get_vm()
|
|
|
|
# VM already exists
|
|
if vm:
|
|
if module.params['state'] == 'absent':
|
|
# destroy it
|
|
if module.params['force']:
|
|
# has to be poweredoff first
|
|
set_vm_power_state(pyv.content, vm, 'poweredoff', module.params['force'])
|
|
result = pyv.remove_vm(vm)
|
|
elif module.params['state'] == 'present':
|
|
result = pyv.reconfigure_vm()
|
|
elif module.params['state'] in ['poweredon', 'poweredoff', 'restarted', 'suspended', 'shutdownguest', 'rebootguest']:
|
|
# set powerstate
|
|
tmp_result = set_vm_power_state(pyv.content, vm, module.params['state'], module.params['force'])
|
|
if tmp_result['changed']:
|
|
result["changed"] = True
|
|
if not tmp_result["failed"]:
|
|
result["failed"] = False
|
|
else:
|
|
# This should not happen
|
|
assert False
|
|
# VM doesn't exist
|
|
else:
|
|
if module.params['state'] in ['poweredon', 'poweredoff', 'present', 'restarted', 'suspended']:
|
|
# Create it ...
|
|
result = pyv.deploy_vm()
|
|
|
|
if result['failed']:
|
|
module.fail_json(**result)
|
|
else:
|
|
module.exit_json(**result)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|