mirror of
https://github.com/ansible-collections/community.general.git
synced 2025-07-24 05:40:23 -07:00
Remove DigitalOcean modules (moved to community.digitalocean) (#622)
* Remove DigitalOcean modules. * Remove inventory scripts.
This commit is contained in:
parent
25aec0d712
commit
c1b5b51366
84 changed files with 154 additions and 5011 deletions
|
@ -1,34 +0,0 @@
|
|||
# Ansible DigitalOcean external inventory script settings
|
||||
#
|
||||
|
||||
[digital_ocean]
|
||||
|
||||
# The module needs your DigitalOcean API Token.
|
||||
# It may also be specified on the command line via --api-token
|
||||
# or via the environment variables DO_API_TOKEN or DO_API_KEY
|
||||
#
|
||||
#api_token = 123456abcdefg
|
||||
|
||||
|
||||
# API calls to DigitalOcean may be slow. For this reason, we cache the results
|
||||
# of an API call. Set this to the path you want cache files to be written to.
|
||||
# One file will be written to this directory:
|
||||
# - ansible-digital_ocean.cache
|
||||
#
|
||||
cache_path = /tmp
|
||||
|
||||
|
||||
# The number of seconds a cache file is considered valid. After this many
|
||||
# seconds, a new API call will be made, and the cache file will be updated.
|
||||
#
|
||||
cache_max_age = 300
|
||||
|
||||
# Use the private network IP address instead of the public when available.
|
||||
#
|
||||
use_private_network = False
|
||||
|
||||
# Pass variables to every group, e.g.:
|
||||
#
|
||||
# group_variables = { 'ansible_user': 'root' }
|
||||
#
|
||||
group_variables = {}
|
|
@ -1,541 +0,0 @@
|
|||
#!/usr/bin/env python
|
||||
|
||||
"""
|
||||
DigitalOcean external inventory script
|
||||
======================================
|
||||
|
||||
Generates Ansible inventory of DigitalOcean Droplets.
|
||||
|
||||
In addition to the --list and --host options used by Ansible, there are options
|
||||
for generating JSON of other DigitalOcean data. This is useful when creating
|
||||
droplets. For example, --regions will return all the DigitalOcean Regions.
|
||||
This information can also be easily found in the cache file, whose default
|
||||
location is /tmp/ansible-digital_ocean.cache).
|
||||
|
||||
The --pretty (-p) option pretty-prints the output for better human readability.
|
||||
|
||||
----
|
||||
Although the cache stores all the information received from DigitalOcean,
|
||||
the cache is not used for current droplet information (in --list, --host,
|
||||
--all, and --droplets). This is so that accurate droplet information is always
|
||||
found. You can force this script to use the cache with --force-cache.
|
||||
|
||||
----
|
||||
Configuration is read from `digital_ocean.ini`, then from environment variables,
|
||||
and then from command-line arguments.
|
||||
|
||||
Most notably, the DigitalOcean API Token must be specified. It can be specified
|
||||
in the INI file or with the following environment variables:
|
||||
export DO_API_TOKEN='abc123' or
|
||||
export DO_API_KEY='abc123'
|
||||
|
||||
Alternatively, it can be passed on the command-line with --api-token.
|
||||
|
||||
If you specify DigitalOcean credentials in the INI file, a handy way to
|
||||
get them into your environment (e.g., to use the digital_ocean module)
|
||||
is to use the output of the --env option with export:
|
||||
export $(digital_ocean.py --env)
|
||||
|
||||
----
|
||||
The following groups are generated from --list:
|
||||
- ID (droplet ID)
|
||||
- NAME (droplet NAME)
|
||||
- digital_ocean
|
||||
- image_ID
|
||||
- image_NAME
|
||||
- distro_NAME (distribution NAME from image)
|
||||
- region_NAME
|
||||
- size_NAME
|
||||
- status_STATUS
|
||||
|
||||
For each host, the following variables are registered:
|
||||
- do_backup_ids
|
||||
- do_created_at
|
||||
- do_disk
|
||||
- do_features - list
|
||||
- do_id
|
||||
- do_image - object
|
||||
- do_ip_address
|
||||
- do_private_ip_address
|
||||
- do_kernel - object
|
||||
- do_locked
|
||||
- do_memory
|
||||
- do_name
|
||||
- do_networks - object
|
||||
- do_next_backup_window
|
||||
- do_region - object
|
||||
- do_size - object
|
||||
- do_size_slug
|
||||
- do_snapshot_ids - list
|
||||
- do_status
|
||||
- do_tags
|
||||
- do_vcpus
|
||||
- do_volume_ids
|
||||
|
||||
-----
|
||||
```
|
||||
usage: digital_ocean.py [-h] [--list] [--host HOST] [--all] [--droplets]
|
||||
[--regions] [--images] [--sizes] [--ssh-keys]
|
||||
[--domains] [--tags] [--pretty]
|
||||
[--cache-path CACHE_PATH]
|
||||
[--cache-max_age CACHE_MAX_AGE] [--force-cache]
|
||||
[--refresh-cache] [--env] [--api-token API_TOKEN]
|
||||
|
||||
Produce an Ansible Inventory file based on DigitalOcean credentials
|
||||
|
||||
optional arguments:
|
||||
-h, --help show this help message and exit
|
||||
--list List all active Droplets as Ansible inventory
|
||||
(default: True)
|
||||
--host HOST Get all Ansible inventory variables about a specific
|
||||
Droplet
|
||||
--all List all DigitalOcean information as JSON
|
||||
--droplets, -d List Droplets as JSON
|
||||
--regions List Regions as JSON
|
||||
--images List Images as JSON
|
||||
--sizes List Sizes as JSON
|
||||
--ssh-keys List SSH keys as JSON
|
||||
--domains List Domains as JSON
|
||||
--tags List Tags as JSON
|
||||
--pretty, -p Pretty-print results
|
||||
--cache-path CACHE_PATH
|
||||
Path to the cache files (default: .)
|
||||
--cache-max_age CACHE_MAX_AGE
|
||||
Maximum age of the cached items (default: 0)
|
||||
--force-cache Only use data from the cache
|
||||
--refresh-cache, -r Force refresh of cache by making API requests to
|
||||
DigitalOcean (default: False - use cache files)
|
||||
--env, -e Display DO_API_TOKEN
|
||||
--api-token API_TOKEN, -a API_TOKEN
|
||||
DigitalOcean API Token
|
||||
```
|
||||
|
||||
"""
|
||||
|
||||
# (c) 2013, Evan Wies <evan@neomantra.net>
|
||||
# (c) 2017, Ansible Project
|
||||
# (c) 2017, Abhijeet Kasurde <akasurde@redhat.com>
|
||||
#
|
||||
# Inspired by the EC2 inventory plugin:
|
||||
# https://github.com/ansible/ansible/blob/devel/contrib/inventory/ec2.py
|
||||
#
|
||||
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
|
||||
|
||||
from __future__ import (absolute_import, division, print_function)
|
||||
__metaclass__ = type
|
||||
|
||||
######################################################################
|
||||
|
||||
import argparse
|
||||
import ast
|
||||
import os
|
||||
import re
|
||||
import requests
|
||||
import sys
|
||||
from time import time
|
||||
|
||||
try:
|
||||
import ConfigParser
|
||||
except ImportError:
|
||||
import configparser as ConfigParser
|
||||
|
||||
import json
|
||||
|
||||
|
||||
class DoManager:
|
||||
def __init__(self, api_token):
|
||||
self.api_token = api_token
|
||||
self.api_endpoint = 'https://api.digitalocean.com/v2'
|
||||
self.headers = {'Authorization': 'Bearer {0}'.format(self.api_token),
|
||||
'Content-type': 'application/json'}
|
||||
self.timeout = 60
|
||||
|
||||
def _url_builder(self, path):
|
||||
if path[0] == '/':
|
||||
path = path[1:]
|
||||
return '%s/%s' % (self.api_endpoint, path)
|
||||
|
||||
def send(self, url, method='GET', data=None):
|
||||
url = self._url_builder(url)
|
||||
data = json.dumps(data)
|
||||
try:
|
||||
if method == 'GET':
|
||||
resp_data = {}
|
||||
incomplete = True
|
||||
while incomplete:
|
||||
resp = requests.get(url, data=data, headers=self.headers, timeout=self.timeout)
|
||||
json_resp = resp.json()
|
||||
|
||||
for key, value in json_resp.items():
|
||||
if isinstance(value, list) and key in resp_data:
|
||||
resp_data[key] += value
|
||||
else:
|
||||
resp_data[key] = value
|
||||
|
||||
try:
|
||||
url = json_resp['links']['pages']['next']
|
||||
except KeyError:
|
||||
incomplete = False
|
||||
|
||||
except ValueError as e:
|
||||
sys.exit("Unable to parse result from %s: %s" % (url, e))
|
||||
return resp_data
|
||||
|
||||
def all_active_droplets(self):
|
||||
resp = self.send('droplets/')
|
||||
return resp['droplets']
|
||||
|
||||
def all_regions(self):
|
||||
resp = self.send('regions/')
|
||||
return resp['regions']
|
||||
|
||||
def all_images(self, filter_name='global'):
|
||||
params = {'filter': filter_name}
|
||||
resp = self.send('images/', data=params)
|
||||
return resp['images']
|
||||
|
||||
def sizes(self):
|
||||
resp = self.send('sizes/')
|
||||
return resp['sizes']
|
||||
|
||||
def all_ssh_keys(self):
|
||||
resp = self.send('account/keys')
|
||||
return resp['ssh_keys']
|
||||
|
||||
def all_domains(self):
|
||||
resp = self.send('domains/')
|
||||
return resp['domains']
|
||||
|
||||
def show_droplet(self, droplet_id):
|
||||
resp = self.send('droplets/%s' % droplet_id)
|
||||
return resp['droplet']
|
||||
|
||||
def all_tags(self):
|
||||
resp = self.send('tags')
|
||||
return resp['tags']
|
||||
|
||||
|
||||
class DigitalOceanInventory(object):
|
||||
|
||||
###########################################################################
|
||||
# Main execution path
|
||||
###########################################################################
|
||||
|
||||
def __init__(self):
|
||||
"""Main execution path """
|
||||
|
||||
# DigitalOceanInventory data
|
||||
self.data = {} # All DigitalOcean data
|
||||
self.inventory = {} # Ansible Inventory
|
||||
|
||||
# Define defaults
|
||||
self.cache_path = '.'
|
||||
self.cache_max_age = 0
|
||||
self.use_private_network = False
|
||||
self.group_variables = {}
|
||||
|
||||
# Read settings, environment variables, and CLI arguments
|
||||
self.read_settings()
|
||||
self.read_environment()
|
||||
self.read_cli_args()
|
||||
|
||||
# Verify credentials were set
|
||||
if not hasattr(self, 'api_token'):
|
||||
msg = 'Could not find values for DigitalOcean api_token. They must be specified via either ini file, ' \
|
||||
'command line argument (--api-token), or environment variables (DO_API_TOKEN)\n'
|
||||
sys.stderr.write(msg)
|
||||
sys.exit(-1)
|
||||
|
||||
# env command, show DigitalOcean credentials
|
||||
if self.args.env:
|
||||
print("DO_API_TOKEN=%s" % self.api_token)
|
||||
sys.exit(0)
|
||||
|
||||
# Manage cache
|
||||
self.cache_filename = self.cache_path + "/ansible-digital_ocean.cache"
|
||||
self.cache_refreshed = False
|
||||
|
||||
if self.is_cache_valid():
|
||||
self.load_from_cache()
|
||||
if len(self.data) == 0:
|
||||
if self.args.force_cache:
|
||||
sys.stderr.write('Cache is empty and --force-cache was specified\n')
|
||||
sys.exit(-1)
|
||||
|
||||
self.manager = DoManager(self.api_token)
|
||||
|
||||
# Pick the json_data to print based on the CLI command
|
||||
if self.args.droplets:
|
||||
self.load_from_digital_ocean('droplets')
|
||||
json_data = {'droplets': self.data['droplets']}
|
||||
elif self.args.regions:
|
||||
self.load_from_digital_ocean('regions')
|
||||
json_data = {'regions': self.data['regions']}
|
||||
elif self.args.images:
|
||||
self.load_from_digital_ocean('images')
|
||||
json_data = {'images': self.data['images']}
|
||||
elif self.args.sizes:
|
||||
self.load_from_digital_ocean('sizes')
|
||||
json_data = {'sizes': self.data['sizes']}
|
||||
elif self.args.ssh_keys:
|
||||
self.load_from_digital_ocean('ssh_keys')
|
||||
json_data = {'ssh_keys': self.data['ssh_keys']}
|
||||
elif self.args.domains:
|
||||
self.load_from_digital_ocean('domains')
|
||||
json_data = {'domains': self.data['domains']}
|
||||
elif self.args.tags:
|
||||
self.load_from_digital_ocean('tags')
|
||||
json_data = {'tags': self.data['tags']}
|
||||
elif self.args.all:
|
||||
self.load_from_digital_ocean()
|
||||
json_data = self.data
|
||||
elif self.args.host:
|
||||
json_data = self.load_droplet_variables_for_host()
|
||||
else: # '--list' this is last to make it default
|
||||
self.load_from_digital_ocean('droplets')
|
||||
self.build_inventory()
|
||||
json_data = self.inventory
|
||||
|
||||
if self.cache_refreshed:
|
||||
self.write_to_cache()
|
||||
|
||||
if self.args.pretty:
|
||||
print(json.dumps(json_data, indent=2))
|
||||
else:
|
||||
print(json.dumps(json_data))
|
||||
|
||||
###########################################################################
|
||||
# Script configuration
|
||||
###########################################################################
|
||||
|
||||
def read_settings(self):
|
||||
""" Reads the settings from the digital_ocean.ini file """
|
||||
config = ConfigParser.ConfigParser()
|
||||
config_path = os.path.join(os.path.dirname(os.path.realpath(__file__)), 'digital_ocean.ini')
|
||||
config.read(config_path)
|
||||
|
||||
# Credentials
|
||||
if config.has_option('digital_ocean', 'api_token'):
|
||||
self.api_token = config.get('digital_ocean', 'api_token')
|
||||
|
||||
# Cache related
|
||||
if config.has_option('digital_ocean', 'cache_path'):
|
||||
self.cache_path = config.get('digital_ocean', 'cache_path')
|
||||
if config.has_option('digital_ocean', 'cache_max_age'):
|
||||
self.cache_max_age = config.getint('digital_ocean', 'cache_max_age')
|
||||
|
||||
# Private IP Address
|
||||
if config.has_option('digital_ocean', 'use_private_network'):
|
||||
self.use_private_network = config.getboolean('digital_ocean', 'use_private_network')
|
||||
|
||||
# Group variables
|
||||
if config.has_option('digital_ocean', 'group_variables'):
|
||||
self.group_variables = ast.literal_eval(config.get('digital_ocean', 'group_variables'))
|
||||
|
||||
def read_environment(self):
|
||||
""" Reads the settings from environment variables """
|
||||
# Setup credentials
|
||||
if os.getenv("DO_API_TOKEN"):
|
||||
self.api_token = os.getenv("DO_API_TOKEN")
|
||||
if os.getenv("DO_API_KEY"):
|
||||
self.api_token = os.getenv("DO_API_KEY")
|
||||
|
||||
def read_cli_args(self):
|
||||
""" Command line argument processing """
|
||||
parser = argparse.ArgumentParser(description='Produce an Ansible Inventory file based on DigitalOcean credentials')
|
||||
|
||||
parser.add_argument('--list', action='store_true', help='List all active Droplets as Ansible inventory (default: True)')
|
||||
parser.add_argument('--host', action='store', help='Get all Ansible inventory variables about a specific Droplet')
|
||||
|
||||
parser.add_argument('--all', action='store_true', help='List all DigitalOcean information as JSON')
|
||||
parser.add_argument('--droplets', '-d', action='store_true', help='List Droplets as JSON')
|
||||
parser.add_argument('--regions', action='store_true', help='List Regions as JSON')
|
||||
parser.add_argument('--images', action='store_true', help='List Images as JSON')
|
||||
parser.add_argument('--sizes', action='store_true', help='List Sizes as JSON')
|
||||
parser.add_argument('--ssh-keys', action='store_true', help='List SSH keys as JSON')
|
||||
parser.add_argument('--domains', action='store_true', help='List Domains as JSON')
|
||||
parser.add_argument('--tags', action='store_true', help='List Tags as JSON')
|
||||
|
||||
parser.add_argument('--pretty', '-p', action='store_true', help='Pretty-print results')
|
||||
|
||||
parser.add_argument('--cache-path', action='store', help='Path to the cache files (default: .)')
|
||||
parser.add_argument('--cache-max_age', action='store', help='Maximum age of the cached items (default: 0)')
|
||||
parser.add_argument('--force-cache', action='store_true', default=False, help='Only use data from the cache')
|
||||
parser.add_argument('--refresh-cache', '-r', action='store_true', default=False,
|
||||
help='Force refresh of cache by making API requests to DigitalOcean (default: False - use cache files)')
|
||||
|
||||
parser.add_argument('--env', '-e', action='store_true', help='Display DO_API_TOKEN')
|
||||
parser.add_argument('--api-token', '-a', action='store', help='DigitalOcean API Token')
|
||||
|
||||
self.args = parser.parse_args()
|
||||
|
||||
if self.args.api_token:
|
||||
self.api_token = self.args.api_token
|
||||
|
||||
# Make --list default if none of the other commands are specified
|
||||
if (not self.args.droplets and not self.args.regions and
|
||||
not self.args.images and not self.args.sizes and
|
||||
not self.args.ssh_keys and not self.args.domains and
|
||||
not self.args.tags and
|
||||
not self.args.all and not self.args.host):
|
||||
self.args.list = True
|
||||
|
||||
###########################################################################
|
||||
# Data Management
|
||||
###########################################################################
|
||||
|
||||
def load_from_digital_ocean(self, resource=None):
|
||||
"""Get JSON from DigitalOcean API """
|
||||
if self.args.force_cache and os.path.isfile(self.cache_filename):
|
||||
return
|
||||
# We always get fresh droplets
|
||||
if self.is_cache_valid() and not (resource == 'droplets' or resource is None):
|
||||
return
|
||||
if self.args.refresh_cache:
|
||||
resource = None
|
||||
|
||||
if resource == 'droplets' or resource is None:
|
||||
self.data['droplets'] = self.manager.all_active_droplets()
|
||||
self.cache_refreshed = True
|
||||
if resource == 'regions' or resource is None:
|
||||
self.data['regions'] = self.manager.all_regions()
|
||||
self.cache_refreshed = True
|
||||
if resource == 'images' or resource is None:
|
||||
self.data['images'] = self.manager.all_images()
|
||||
self.cache_refreshed = True
|
||||
if resource == 'sizes' or resource is None:
|
||||
self.data['sizes'] = self.manager.sizes()
|
||||
self.cache_refreshed = True
|
||||
if resource == 'ssh_keys' or resource is None:
|
||||
self.data['ssh_keys'] = self.manager.all_ssh_keys()
|
||||
self.cache_refreshed = True
|
||||
if resource == 'domains' or resource is None:
|
||||
self.data['domains'] = self.manager.all_domains()
|
||||
self.cache_refreshed = True
|
||||
if resource == 'tags' or resource is None:
|
||||
self.data['tags'] = self.manager.all_tags()
|
||||
self.cache_refreshed = True
|
||||
|
||||
def add_inventory_group(self, key):
|
||||
""" Method to create group dict """
|
||||
host_dict = {'hosts': [], 'vars': {}}
|
||||
self.inventory[key] = host_dict
|
||||
return
|
||||
|
||||
def add_host(self, group, host):
|
||||
""" Helper method to reduce host duplication """
|
||||
if group not in self.inventory:
|
||||
self.add_inventory_group(group)
|
||||
|
||||
if host not in self.inventory[group]['hosts']:
|
||||
self.inventory[group]['hosts'].append(host)
|
||||
return
|
||||
|
||||
def build_inventory(self):
|
||||
""" Build Ansible inventory of droplets """
|
||||
self.inventory = {
|
||||
'all': {
|
||||
'hosts': [],
|
||||
'vars': self.group_variables
|
||||
},
|
||||
'_meta': {'hostvars': {}}
|
||||
}
|
||||
|
||||
# add all droplets by id and name
|
||||
for droplet in self.data['droplets']:
|
||||
for net in droplet['networks']['v4']:
|
||||
if net['type'] == 'public':
|
||||
dest = net['ip_address']
|
||||
else:
|
||||
continue
|
||||
|
||||
self.inventory['all']['hosts'].append(dest)
|
||||
|
||||
self.add_host(droplet['id'], dest)
|
||||
|
||||
self.add_host(droplet['name'], dest)
|
||||
|
||||
# groups that are always present
|
||||
for group in ('digital_ocean',
|
||||
'region_' + droplet['region']['slug'],
|
||||
'image_' + str(droplet['image']['id']),
|
||||
'size_' + droplet['size']['slug'],
|
||||
'distro_' + DigitalOceanInventory.to_safe(droplet['image']['distribution']),
|
||||
'status_' + droplet['status']):
|
||||
self.add_host(group, dest)
|
||||
|
||||
# groups that are not always present
|
||||
for group in (droplet['image']['slug'],
|
||||
droplet['image']['name']):
|
||||
if group:
|
||||
image = 'image_' + DigitalOceanInventory.to_safe(group)
|
||||
self.add_host(image, dest)
|
||||
|
||||
if droplet['tags']:
|
||||
for tag in droplet['tags']:
|
||||
self.add_host(tag, dest)
|
||||
|
||||
# hostvars
|
||||
info = self.do_namespace(droplet)
|
||||
self.inventory['_meta']['hostvars'][dest] = info
|
||||
|
||||
def load_droplet_variables_for_host(self):
|
||||
""" Generate a JSON response to a --host call """
|
||||
host = int(self.args.host)
|
||||
droplet = self.manager.show_droplet(host)
|
||||
info = self.do_namespace(droplet)
|
||||
return {'droplet': info}
|
||||
|
||||
###########################################################################
|
||||
# Cache Management
|
||||
###########################################################################
|
||||
|
||||
def is_cache_valid(self):
|
||||
""" Determines if the cache files have expired, or if it is still valid """
|
||||
if os.path.isfile(self.cache_filename):
|
||||
mod_time = os.path.getmtime(self.cache_filename)
|
||||
current_time = time()
|
||||
if (mod_time + self.cache_max_age) > current_time:
|
||||
return True
|
||||
return False
|
||||
|
||||
def load_from_cache(self):
|
||||
""" Reads the data from the cache file and assigns it to member variables as Python Objects """
|
||||
try:
|
||||
with open(self.cache_filename, 'r') as cache:
|
||||
json_data = cache.read()
|
||||
data = json.loads(json_data)
|
||||
except IOError:
|
||||
data = {'data': {}, 'inventory': {}}
|
||||
|
||||
self.data = data['data']
|
||||
self.inventory = data['inventory']
|
||||
|
||||
def write_to_cache(self):
|
||||
""" Writes data in JSON format to a file """
|
||||
data = {'data': self.data, 'inventory': self.inventory}
|
||||
json_data = json.dumps(data, indent=2)
|
||||
|
||||
with open(self.cache_filename, 'w') as cache:
|
||||
cache.write(json_data)
|
||||
|
||||
###########################################################################
|
||||
# Utilities
|
||||
###########################################################################
|
||||
@staticmethod
|
||||
def to_safe(word):
|
||||
""" Converts 'bad' characters in a string to underscores so they can be used as Ansible groups """
|
||||
return re.sub(r"[^A-Za-z0-9\-.]", "_", word)
|
||||
|
||||
@staticmethod
|
||||
def do_namespace(data):
|
||||
""" Returns a copy of the dictionary with all the keys put in a 'do_' namespace """
|
||||
info = {}
|
||||
for k, v in data.items():
|
||||
info['do_' + k] = v
|
||||
return info
|
||||
|
||||
|
||||
###########################################################################
|
||||
# Run the script
|
||||
DigitalOceanInventory()
|
Loading…
Add table
Add a link
Reference in a new issue