google.cloud/plugins/modules/gcp_cloudbuild_trigger.py
2024-11-05 14:06:56 -08:00

2256 lines
86 KiB
Python

#!/usr/bin/python
# -*- coding: utf-8 -*-
#
# Copyright (C) 2017 Google
# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt)
# ----------------------------------------------------------------------------
#
# *** AUTO GENERATED CODE *** Type: MMv1 ***
#
# ----------------------------------------------------------------------------
#
# This file is automatically generated by Magic Modules and manual
# changes will be clobbered when the file is regenerated.
#
# Please read more about how to change this file at
# https://www.github.com/GoogleCloudPlatform/magic-modules
#
# ----------------------------------------------------------------------------
from __future__ import absolute_import, division, print_function
__metaclass__ = type
################################################################################
# Documentation
################################################################################
ANSIBLE_METADATA = {'metadata_version': '1.1', 'status': ["preview"], 'supported_by': 'community'}
DOCUMENTATION = '''
---
module: gcp_cloudbuild_trigger
description:
- Configuration for an automated build in response to source repository changes.
short_description: Creates a GCP Trigger
author: Google Inc. (@googlecloudplatform)
requirements:
- python >= 2.6
- requests >= 2.18.4
- google-auth >= 1.3.0
options:
state:
description:
- Whether the given object should exist in GCP
choices:
- present
- absent
default: present
type: str
id:
description:
- The unique identifier for the trigger.
required: false
type: str
name:
description:
- Name of the trigger. Must be unique within the project.
required: false
type: str
description:
description:
- Human-readable description of the trigger.
required: false
type: str
tags:
description:
- Tags for annotation of a BuildTrigger .
elements: str
required: false
type: list
disabled:
description:
- Whether the trigger is disabled or not. If true, the trigger will never result
in a build.
required: false
type: bool
substitutions:
description:
- Substitutions data for Build resource.
required: false
type: dict
filename:
description:
- Path, from the source root, to a file whose contents is used for the template.
Either a filename or build template must be provided.
required: false
type: str
ignored_files:
description:
- ignoredFiles and includedFiles are file glob matches using U(https://golang.org/pkg/path/filepath/#Match)
extended with support for `**`.
- If ignoredFiles and changed files are both empty, then they are not used to
determine whether or not to trigger a build.
- If ignoredFiles is not empty, then we ignore any files that match any of the
ignored_file globs. If the change has no files that are outside of the ignoredFiles
globs, then we do not trigger a build.
elements: str
required: false
type: list
included_files:
description:
- ignoredFiles and includedFiles are file glob matches using U(https://golang.org/pkg/path/filepath/#Match)
extended with support for `**`.
- If any of the files altered in the commit pass the ignoredFiles filter and includedFiles
is empty, then as far as this filter is concerned, we should trigger the build.
- If any of the files altered in the commit pass the ignoredFiles filter and includedFiles
is not empty, then we make sure that at least one of those files matches a includedFiles
glob. If not, then we do not trigger a build.
elements: str
required: false
type: list
trigger_template:
description:
- Template describing the types of source changes to trigger a build.
- Branch and tag names in trigger templates are interpreted as regular expressions.
Any branch or tag change that matches that regular expression will trigger a
build.
required: false
type: dict
suboptions:
project_id:
description:
- ID of the project that owns the Cloud Source Repository. If omitted, the
project ID requesting the build is assumed.
required: false
type: str
repo_name:
description:
- Name of the Cloud Source Repository. If omitted, the name "default" is assumed.
required: false
default: default
type: str
dir:
description:
- Directory, relative to the source root, in which to run the build.
- This must be a relative path. If a step's dir is specified and is an absolute
path, this value is ignored for that step's execution.
required: false
type: str
invert_regex:
description:
- Only trigger a build if the revision regex does NOT match the revision regex.
required: false
type: bool
branch_name:
description:
- Name of the branch to build. Exactly one a of branch name, tag, or commit
SHA must be provided.
- This field is a regular expression.
required: false
type: str
tag_name:
description:
- Name of the tag to build. Exactly one of a branch name, tag, or commit SHA
must be provided.
- This field is a regular expression.
required: false
type: str
commit_sha:
description:
- Explicit commit SHA to build. Exactly one of a branch name, tag, or commit
SHA must be provided.
required: false
type: str
github:
description:
- Describes the configuration of a trigger that creates a build whenever a GitHub
event is received.
required: false
type: dict
suboptions:
owner:
description:
- 'Owner of the repository. For example: The owner for U(https://github.com/googlecloudplatform/cloud-builders)
is "googlecloudplatform".'
required: false
type: str
name:
description:
- 'Name of the repository. For example: The name for U(https://github.com/googlecloudplatform/cloud-builders)
is "cloud-builders".'
required: false
type: str
pull_request:
description:
- filter to match changes in pull requests. Specify only one of pullRequest
or push.
required: false
type: dict
suboptions:
branch:
description:
- Regex of branches to match.
required: true
type: str
comment_control:
description:
- Whether to block builds on a "/gcbrun" comment from a repository owner
or collaborator.
- 'Some valid choices include: "COMMENTS_DISABLED", "COMMENTS_ENABLED",
"COMMENTS_ENABLED_FOR_EXTERNAL_CONTRIBUTORS_ONLY"'
required: false
type: str
invert_regex:
description:
- If true, branches that do NOT match the git_ref will trigger a build.
required: false
type: bool
push:
description:
- filter to match changes in refs, like branches or tags. Specify only one
of pullRequest or push.
required: false
type: dict
suboptions:
invert_regex:
description:
- When true, only trigger a build if the revision regex does NOT match
the git_ref regex.
required: false
type: bool
branch:
description:
- Regex of branches to match. Specify only one of branch or tag.
required: false
type: str
tag:
description:
- Regex of tags to match. Specify only one of branch or tag.
required: false
type: str
pubsub_config:
description:
- PubsubConfig describes the configuration of a trigger that creates a build whenever
a Pub/Sub message is published.
required: false
type: dict
suboptions:
topic:
description:
- The name of the topic from which this subscription is receiving messages.
required: true
type: str
service_account_email:
description:
- Service account that will make the push request.
required: false
type: str
webhook_config:
description:
- WebhookConfig describes the configuration of a trigger that creates a build
whenever a webhook is sent to a trigger's webhook URL.
required: false
type: dict
suboptions:
secret:
description:
- Resource name for the secret required as a URL parameter.
required: true
type: str
build:
description:
- Contents of the build template. Either a filename or build template must be
provided.
required: false
type: dict
suboptions:
source:
description:
- The location of the source files to build.
required: false
type: dict
suboptions:
storage_source:
description:
- Location of the source in an archive file in Google Cloud Storage.
required: false
type: dict
suboptions:
bucket:
description:
- Google Cloud Storage bucket containing the source.
required: true
type: str
object:
description:
- Google Cloud Storage object containing the source.
- This object must be a gzipped archive file (.tar.gz) containing
source to build.
required: true
type: str
generation:
description:
- Google Cloud Storage generation for the object. If the generation
is omitted, the latest generation will be used .
required: false
type: str
repo_source:
description:
- Location of the source in a Google Cloud Source Repository.
required: false
type: dict
suboptions:
project_id:
description:
- ID of the project that owns the Cloud Source Repository. If omitted,
the project ID requesting the build is assumed.
required: false
type: str
repo_name:
description:
- Name of the Cloud Source Repository.
required: true
type: str
dir:
description:
- Directory, relative to the source root, in which to run the build.
- This must be a relative path. If a step's dir is specified and is
an absolute path, this value is ignored for that step's execution.
required: false
type: str
invert_regex:
description:
- Only trigger a build if the revision regex does NOT match the revision
regex.
required: false
type: bool
substitutions:
description:
- Substitutions to use in a triggered build. Should only be used with
triggers.run .
required: false
type: dict
branch_name:
description:
- Regex matching branches to build. Exactly one a of branch name,
tag, or commit SHA must be provided.
- The syntax of the regular expressions accepted is the syntax accepted
by RE2 and described at U(https://github.com/google/re2/wiki/Syntax)
.
required: false
type: str
tag_name:
description:
- Regex matching tags to build. Exactly one a of branch name, tag,
or commit SHA must be provided.
- The syntax of the regular expressions accepted is the syntax accepted
by RE2 and described at U(https://github.com/google/re2/wiki/Syntax)
.
required: false
type: str
commit_sha:
description:
- Explicit commit SHA to build. Exactly one a of branch name, tag,
or commit SHA must be provided.
required: false
type: str
tags:
description:
- Tags for annotation of a Build. These are not docker tags.
elements: str
required: false
type: list
images:
description:
- A list of images to be pushed upon the successful completion of all build
steps.
- The images are pushed using the builder service account's credentials.
- The digests of the pushed images will be stored in the Build resource's
results field.
- If any of the images fail to be pushed, the build status is marked FAILURE.
elements: str
required: false
type: list
substitutions:
description:
- Substitutions data for Build resource.
required: false
type: dict
queue_ttl:
description:
- TTL in queue for this build. If provided and the build is enqueued longer
than this value, the build will expire and the build status will be EXPIRED.
- The TTL starts ticking from createTime.
- 'A duration in seconds with up to nine fractional digits, terminated by
''s''. Example: "3.5s".'
required: false
type: str
logs_bucket:
description:
- Google Cloud Storage bucket where logs should be written. Logs file names
will be of the format ${logsBucket}/log-${build_id}.txt.
required: false
type: str
timeout:
description:
- Amount of time that this build should be allowed to run, to second granularity.
- If this amount of time elapses, work on the build will cease and the build
status will be TIMEOUT.
- This timeout must be equal to or greater than the sum of the timeouts for
build steps within the build.
- The expected format is the number of seconds followed by s.
- Default time is ten minutes (600s).
required: false
default: 600s
type: str
secrets:
description:
- Secrets to decrypt using Cloud Key Management Service.
elements: dict
required: false
type: list
suboptions:
kms_key_name:
description:
- Cloud KMS key name to use to decrypt these envs.
required: true
type: str
secret_env:
description:
- Map of environment variable name to its encrypted value.
- Secret environment variables must be unique across all of a build's
secrets, and must be used by at least one build step. Values can be
at most 64 KB in size. There can be at most 100 secret values across
all of a build's secrets.
required: false
type: dict
steps:
description:
- The operations to be performed on the workspace.
elements: dict
required: true
type: list
suboptions:
name:
description:
- The name of the container image that will run this particular build
step.
- If the image is available in the host's Docker daemon's cache, it will
be run directly. If not, the host will attempt to pull the image first,
using the builder service account's credentials if necessary.
- The Docker daemon's cache will already have the latest versions of all
of the officially supported build steps (see U(https://github.com/GoogleCloudPlatform/cloud-builders)
for images and examples).
- The Docker daemon will also have cached many of the layers for some
popular images, like "ubuntu", "debian", but they will be refreshed
at the time you attempt to use them.
- If you built an image in a previous build step, it will be stored in
the host's Docker daemon's cache and is available to use as the name
for a later build step.
required: true
type: str
args:
description:
- A list of arguments that will be presented to the step when it is started.
- If the image used to run the step's container has an entrypoint, the
args are used as arguments to that entrypoint. If the image does not
define an entrypoint, the first element in args is used as the entrypoint,
and the remainder will be used as arguments.
elements: str
required: false
type: list
env:
description:
- A list of environment variable definitions to be used when running a
step.
- The elements are of the form "KEY=VALUE" for the environment variable
"KEY" being given the value "VALUE".
elements: str
required: false
type: list
id:
description:
- Unique identifier for this build step, used in `wait_for` to reference
this build step as a dependency.
required: false
type: str
entrypoint:
description:
- Entrypoint to be used instead of the build step image's default entrypoint.
- If unset, the image's default entrypoint is used .
required: false
type: str
dir:
description:
- Working directory to use when running this step's container.
- If this value is a relative path, it is relative to the build's working
directory. If this value is absolute, it may be outside the build's
working directory, in which case the contents of the path may not be
persisted across build step executions, unless a `volume` for that path
is specified.
- If the build specifies a `RepoSource` with `dir` and a step with a `dir`,
which specifies an absolute path, the `RepoSource` `dir` is ignored
for the step's execution.
required: false
type: str
secret_env:
description:
- A list of environment variables which are encrypted using a Cloud Key
Management Service crypto key. These values must be specified in the
build's `Secret`.
elements: str
required: false
type: list
timeout:
description:
- Time limit for executing this build step. If not defined, the step has
no time limit and will be allowed to continue to run until either it
completes or the build itself times out.
required: false
type: str
timing:
description:
- Output only. Stores timing information for executing this build step.
required: false
type: str
volumes:
description:
- List of volumes to mount into the build step.
- Each volume is created as an empty volume prior to execution of the
build step. Upon completion of the build, volumes and their contents
are discarded.
- Using a named volume in only one step is not valid as it is indicative
of a build request with an incorrect configuration.
elements: dict
required: false
type: list
suboptions:
name:
description:
- Name of the volume to mount.
- Volume names must be unique per build step and must be valid names
for Docker volumes. Each named volume must be used by at least two
build steps.
required: true
type: str
path:
description:
- Path at which to mount the volume.
- Paths must be absolute and cannot conflict with other volume paths
on the same build step or with certain reserved volume paths.
required: true
type: str
wait_for:
description:
- The ID(s) of the step(s) that this build step depends on.
- This build step will not start until all the build steps in `wait_for`
have completed successfully. If `wait_for` is empty, this build step
will start when all previous build steps in the `Build.Steps` list have
completed successfully.
elements: str
required: false
type: list
artifacts:
description:
- Artifacts produced by the build that should be uploaded upon successful
completion of all build steps.
required: false
type: dict
suboptions:
images:
description:
- A list of images to be pushed upon the successful completion of all
build steps.
- The images will be pushed using the builder service account's credentials.
- The digests of the pushed images will be stored in the Build resource's
results field.
- If any of the images fail to be pushed, the build is marked FAILURE.
elements: str
required: false
type: list
objects:
description:
- A list of objects to be uploaded to Cloud Storage upon successful completion
of all build steps.
- Files in the workspace matching specified paths globs will be uploaded
to the Cloud Storage location using the builder service account's credentials.
- The location and generation of the uploaded objects will be stored in
the Build resource's results field.
- If any objects fail to be pushed, the build is marked FAILURE.
required: false
type: dict
suboptions:
location:
description:
- Cloud Storage bucket and optional object path, in the form "gs://bucket/path/to/somewhere/".
- Files in the workspace matching any path pattern will be uploaded
to Cloud Storage with this location as a prefix.
required: false
type: str
paths:
description:
- Path globs used to match files in the build's workspace.
elements: str
required: false
type: list
options:
description:
- Special options for this build.
required: false
type: dict
suboptions:
source_provenance_hash:
description:
- Requested hash for SourceProvenance.
elements: str
required: false
type: list
requested_verify_option:
description:
- Requested verifiability options.
- 'Some valid choices include: "NOT_VERIFIED", "VERIFIED"'
required: false
type: str
machine_type:
description:
- Compute Engine machine type on which to run the build.
- 'Some valid choices include: "UNSPECIFIED", "N1_HIGHCPU_8", "N1_HIGHCPU_32",
"E2_HIGHCPU_8", "E2_HIGHCPU_32"'
required: false
type: str
disk_size_gb:
description:
- Requested disk size for the VM that runs the build. Note that this is
NOT "disk free"; some of the space will be used by the operating system
and build utilities.
- Also note that this is the minimum disk size that will be allocated
for the build -- the build may run with a larger disk than requested.
At present, the maximum disk size is 1000GB; builds that request more
than the maximum are rejected with an error.
required: false
type: int
substitution_option:
description:
- Option to specify behavior when there is an error in the substitution
checks.
- NOTE this is always set to ALLOW_LOOSE for triggered builds and cannot
be overridden in the build configuration file.
- 'Some valid choices include: "MUST_MATCH", "ALLOW_LOOSE"'
required: false
type: str
dynamic_substitutions:
description:
- Option to specify whether or not to apply bash style string operations
to the substitutions.
- NOTE this is always enabled for triggered builds and cannot be overridden
in the build configuration file.
required: false
type: bool
log_streaming_option:
description:
- Option to define build log streaming behavior to Google Cloud Storage.
- 'Some valid choices include: "STREAM_DEFAULT", "STREAM_ON", "STREAM_OFF"'
required: false
type: str
worker_pool:
description:
- Option to specify a WorkerPool for the build. Format projects/{project}/workerPools/{workerPool}
This field is experimental.
required: false
type: str
logging:
description:
- Option to specify the logging mode, which determines if and where build
logs are stored.
- 'Some valid choices include: "LOGGING_UNSPECIFIED", "LEGACY", "GCS_ONLY",
"STACKDRIVER_ONLY", "NONE"'
required: false
type: str
env:
description:
- A list of global environment variable definitions that will exist for
all build steps in this build. If a variable is defined in both globally
and in a build step, the variable will use the build step value.
- The elements are of the form "KEY=VALUE" for the environment variable
"KEY" being given the value "VALUE".
elements: str
required: false
type: list
secret_env:
description:
- A list of global environment variables, which are encrypted using a
Cloud Key Management Service crypto key. These values must be specified
in the build's Secret. These variables will be available to all build
steps in this build.
elements: str
required: false
type: list
volumes:
description:
- Global list of volumes to mount for ALL build steps Each volume is created
as an empty volume prior to starting the build process.
- Upon completion of the build, volumes and their contents are discarded.
Global volume names and paths cannot conflict with the volumes defined
a build step.
- Using a global volume in a build with only one step is not valid as
it is indicative of a build request with an incorrect configuration.
elements: dict
required: false
type: list
suboptions:
name:
description:
- Name of the volume to mount.
- Volume names must be unique per build step and must be valid names
for Docker volumes.
- Each named volume must be used by at least two build steps.
required: false
type: str
path:
description:
- Path at which to mount the volume.
- Paths must be absolute and cannot conflict with other volume paths
on the same build step or with certain reserved volume paths.
required: false
type: str
project:
description:
- The Google Cloud Platform project to use.
type: str
auth_kind:
description:
- The type of credential used.
type: str
required: true
choices:
- application
- machineaccount
- serviceaccount
- accesstoken
service_account_contents:
description:
- The contents of a Service Account JSON file, either in a dictionary or as a
JSON string that represents it.
type: jsonarg
service_account_file:
description:
- The path of a Service Account JSON file if serviceaccount is selected as type.
type: path
service_account_email:
description:
- An optional service account email address if machineaccount is selected and
the user does not wish to use the default email.
type: str
access_token:
description:
- An OAuth2 access token if credential type is accesstoken.
type: str
scopes:
description:
- Array of scopes to be used
type: list
elements: str
env_type:
description:
- Specifies which Ansible environment you're running this module within.
- This should not be set unless you know what you're doing.
- This only alters the User Agent string for any API requests.
type: str
notes:
- 'API Reference: U(https://cloud.google.com/cloud-build/docs/api/reference/rest/v1/projects.triggers)'
- 'Automating builds using build triggers: U(https://cloud.google.com/cloud-build/docs/running-builds/automate-builds)'
- for authentication, you can set service_account_file using the C(GCP_SERVICE_ACCOUNT_FILE)
env variable.
- for authentication, you can set service_account_contents using the C(GCP_SERVICE_ACCOUNT_CONTENTS)
env variable.
- For authentication, you can set service_account_email using the C(GCP_SERVICE_ACCOUNT_EMAIL)
env variable.
- For authentication, you can set access_token using the C(GCP_ACCESS_TOKEN)
env variable.
- For authentication, you can set auth_kind using the C(GCP_AUTH_KIND) env variable.
- For authentication, you can set scopes using the C(GCP_SCOPES) env variable.
- Environment variables values will only be used if the playbook values are not set.
- The I(service_account_email) and I(service_account_file) options are mutually exclusive.
- The id for this resource is created by the API after you create the resource the
first time. If you want to manage this resource after creation, you'll have to copy
the generated id into the playbook. If you do not, new triggers will be created
on subsequent runs.
'''
EXAMPLES = '''
- name: create a repository
google.cloud.gcp_sourcerepo_repository:
name: projects/{{ gcp_project }}/repos/{{ resource_name }}
project: "{{ gcp_project }}"
auth_kind: "{{ gcp_cred_kind }}"
service_account_file: "{{ gcp_cred_file }}"
state: present
- name: create a trigger
google.cloud.gcp_cloudbuild_trigger:
trigger_template:
branch_name: master
project_id: test_project
repo_name: test_object
filename: cloudbuild.yaml
project: test_project
auth_kind: serviceaccount
service_account_file: "/tmp/auth.pem"
state: present
'''
RETURN = '''
id:
description:
- The unique identifier for the trigger.
returned: success
type: str
name:
description:
- Name of the trigger. Must be unique within the project.
returned: success
type: str
description:
description:
- Human-readable description of the trigger.
returned: success
type: str
tags:
description:
- Tags for annotation of a BuildTrigger .
returned: success
type: list
disabled:
description:
- Whether the trigger is disabled or not. If true, the trigger will never result
in a build.
returned: success
type: bool
createTime:
description:
- Time when the trigger was created.
returned: success
type: str
substitutions:
description:
- Substitutions data for Build resource.
returned: success
type: dict
filename:
description:
- Path, from the source root, to a file whose contents is used for the template.
Either a filename or build template must be provided.
returned: success
type: str
ignoredFiles:
description:
- ignoredFiles and includedFiles are file glob matches using U(https://golang.org/pkg/path/filepath/#Match)
extended with support for `**`.
- If ignoredFiles and changed files are both empty, then they are not used to determine
whether or not to trigger a build.
- If ignoredFiles is not empty, then we ignore any files that match any of the ignored_file
globs. If the change has no files that are outside of the ignoredFiles globs,
then we do not trigger a build.
returned: success
type: list
includedFiles:
description:
- ignoredFiles and includedFiles are file glob matches using U(https://golang.org/pkg/path/filepath/#Match)
extended with support for `**`.
- If any of the files altered in the commit pass the ignoredFiles filter and includedFiles
is empty, then as far as this filter is concerned, we should trigger the build.
- If any of the files altered in the commit pass the ignoredFiles filter and includedFiles
is not empty, then we make sure that at least one of those files matches a includedFiles
glob. If not, then we do not trigger a build.
returned: success
type: list
triggerTemplate:
description:
- Template describing the types of source changes to trigger a build.
- Branch and tag names in trigger templates are interpreted as regular expressions.
Any branch or tag change that matches that regular expression will trigger a build.
returned: success
type: complex
contains:
projectId:
description:
- ID of the project that owns the Cloud Source Repository. If omitted, the project
ID requesting the build is assumed.
returned: success
type: str
repoName:
description:
- Name of the Cloud Source Repository. If omitted, the name "default" is assumed.
returned: success
type: str
dir:
description:
- Directory, relative to the source root, in which to run the build.
- This must be a relative path. If a step's dir is specified and is an absolute
path, this value is ignored for that step's execution.
returned: success
type: str
invertRegex:
description:
- Only trigger a build if the revision regex does NOT match the revision regex.
returned: success
type: bool
branchName:
description:
- Name of the branch to build. Exactly one a of branch name, tag, or commit
SHA must be provided.
- This field is a regular expression.
returned: success
type: str
tagName:
description:
- Name of the tag to build. Exactly one of a branch name, tag, or commit SHA
must be provided.
- This field is a regular expression.
returned: success
type: str
commitSha:
description:
- Explicit commit SHA to build. Exactly one of a branch name, tag, or commit
SHA must be provided.
returned: success
type: str
github:
description:
- Describes the configuration of a trigger that creates a build whenever a GitHub
event is received.
returned: success
type: complex
contains:
owner:
description:
- 'Owner of the repository. For example: The owner for U(https://github.com/googlecloudplatform/cloud-builders)
is "googlecloudplatform".'
returned: success
type: str
name:
description:
- 'Name of the repository. For example: The name for U(https://github.com/googlecloudplatform/cloud-builders)
is "cloud-builders".'
returned: success
type: str
pullRequest:
description:
- filter to match changes in pull requests. Specify only one of pullRequest
or push.
returned: success
type: complex
contains:
branch:
description:
- Regex of branches to match.
returned: success
type: str
commentControl:
description:
- Whether to block builds on a "/gcbrun" comment from a repository owner
or collaborator.
returned: success
type: str
invertRegex:
description:
- If true, branches that do NOT match the git_ref will trigger a build.
returned: success
type: bool
push:
description:
- filter to match changes in refs, like branches or tags. Specify only one of
pullRequest or push.
returned: success
type: complex
contains:
invertRegex:
description:
- When true, only trigger a build if the revision regex does NOT match the
git_ref regex.
returned: success
type: bool
branch:
description:
- Regex of branches to match. Specify only one of branch or tag.
returned: success
type: str
tag:
description:
- Regex of tags to match. Specify only one of branch or tag.
returned: success
type: str
pubsubConfig:
description:
- PubsubConfig describes the configuration of a trigger that creates a build whenever
a Pub/Sub message is published.
returned: success
type: complex
contains:
subscription:
description:
- Output only. Name of the subscription.
returned: success
type: str
topic:
description:
- The name of the topic from which this subscription is receiving messages.
returned: success
type: str
service_account_email:
description:
- Service account that will make the push request.
returned: success
type: str
state:
description:
- Potential issues with the underlying Pub/Sub subscription configuration.
- Only populated on get requests.
returned: success
type: str
webhookConfig:
description:
- WebhookConfig describes the configuration of a trigger that creates a build whenever
a webhook is sent to a trigger's webhook URL.
returned: success
type: complex
contains:
secret:
description:
- Resource name for the secret required as a URL parameter.
returned: success
type: str
state:
description:
- Potential issues with the underlying Pub/Sub subscription configuration.
- Only populated on get requests.
returned: success
type: str
build:
description:
- Contents of the build template. Either a filename or build template must be provided.
returned: success
type: complex
contains:
source:
description:
- The location of the source files to build.
returned: success
type: complex
contains:
storageSource:
description:
- Location of the source in an archive file in Google Cloud Storage.
returned: success
type: complex
contains:
bucket:
description:
- Google Cloud Storage bucket containing the source.
returned: success
type: str
object:
description:
- Google Cloud Storage object containing the source.
- This object must be a gzipped archive file (.tar.gz) containing source
to build.
returned: success
type: str
generation:
description:
- Google Cloud Storage generation for the object. If the generation
is omitted, the latest generation will be used .
returned: success
type: str
repoSource:
description:
- Location of the source in a Google Cloud Source Repository.
returned: success
type: complex
contains:
projectId:
description:
- ID of the project that owns the Cloud Source Repository. If omitted,
the project ID requesting the build is assumed.
returned: success
type: str
repoName:
description:
- Name of the Cloud Source Repository.
returned: success
type: str
dir:
description:
- Directory, relative to the source root, in which to run the build.
- This must be a relative path. If a step's dir is specified and is
an absolute path, this value is ignored for that step's execution.
returned: success
type: str
invertRegex:
description:
- Only trigger a build if the revision regex does NOT match the revision
regex.
returned: success
type: bool
substitutions:
description:
- Substitutions to use in a triggered build. Should only be used with
triggers.run .
returned: success
type: dict
branchName:
description:
- Regex matching branches to build. Exactly one a of branch name, tag,
or commit SHA must be provided.
- The syntax of the regular expressions accepted is the syntax accepted
by RE2 and described at U(https://github.com/google/re2/wiki/Syntax)
.
returned: success
type: str
tagName:
description:
- Regex matching tags to build. Exactly one a of branch name, tag, or
commit SHA must be provided.
- The syntax of the regular expressions accepted is the syntax accepted
by RE2 and described at U(https://github.com/google/re2/wiki/Syntax)
.
returned: success
type: str
commitSha:
description:
- Explicit commit SHA to build. Exactly one a of branch name, tag, or
commit SHA must be provided.
returned: success
type: str
tags:
description:
- Tags for annotation of a Build. These are not docker tags.
returned: success
type: list
images:
description:
- A list of images to be pushed upon the successful completion of all build
steps.
- The images are pushed using the builder service account's credentials.
- The digests of the pushed images will be stored in the Build resource's results
field.
- If any of the images fail to be pushed, the build status is marked FAILURE.
returned: success
type: list
substitutions:
description:
- Substitutions data for Build resource.
returned: success
type: dict
queueTtl:
description:
- TTL in queue for this build. If provided and the build is enqueued longer
than this value, the build will expire and the build status will be EXPIRED.
- The TTL starts ticking from createTime.
- 'A duration in seconds with up to nine fractional digits, terminated by ''s''.
Example: "3.5s".'
returned: success
type: str
logsBucket:
description:
- Google Cloud Storage bucket where logs should be written. Logs file names
will be of the format ${logsBucket}/log-${build_id}.txt.
returned: success
type: str
timeout:
description:
- Amount of time that this build should be allowed to run, to second granularity.
- If this amount of time elapses, work on the build will cease and the build
status will be TIMEOUT.
- This timeout must be equal to or greater than the sum of the timeouts for
build steps within the build.
- The expected format is the number of seconds followed by s.
- Default time is ten minutes (600s).
returned: success
type: str
secrets:
description:
- Secrets to decrypt using Cloud Key Management Service.
returned: success
type: complex
contains:
kmsKeyName:
description:
- Cloud KMS key name to use to decrypt these envs.
returned: success
type: str
secretEnv:
description:
- Map of environment variable name to its encrypted value.
- Secret environment variables must be unique across all of a build's secrets,
and must be used by at least one build step. Values can be at most 64
KB in size. There can be at most 100 secret values across all of a build's
secrets.
returned: success
type: dict
steps:
description:
- The operations to be performed on the workspace.
returned: success
type: complex
contains:
name:
description:
- The name of the container image that will run this particular build step.
- If the image is available in the host's Docker daemon's cache, it will
be run directly. If not, the host will attempt to pull the image first,
using the builder service account's credentials if necessary.
- The Docker daemon's cache will already have the latest versions of all
of the officially supported build steps (see U(https://github.com/GoogleCloudPlatform/cloud-builders)
for images and examples).
- The Docker daemon will also have cached many of the layers for some popular
images, like "ubuntu", "debian", but they will be refreshed at the time
you attempt to use them.
- If you built an image in a previous build step, it will be stored in the
host's Docker daemon's cache and is available to use as the name for a
later build step.
returned: success
type: str
args:
description:
- A list of arguments that will be presented to the step when it is started.
- If the image used to run the step's container has an entrypoint, the args
are used as arguments to that entrypoint. If the image does not define
an entrypoint, the first element in args is used as the entrypoint, and
the remainder will be used as arguments.
returned: success
type: list
env:
description:
- A list of environment variable definitions to be used when running a step.
- The elements are of the form "KEY=VALUE" for the environment variable
"KEY" being given the value "VALUE".
returned: success
type: list
id:
description:
- Unique identifier for this build step, used in `wait_for` to reference
this build step as a dependency.
returned: success
type: str
entrypoint:
description:
- Entrypoint to be used instead of the build step image's default entrypoint.
- If unset, the image's default entrypoint is used .
returned: success
type: str
dir:
description:
- Working directory to use when running this step's container.
- If this value is a relative path, it is relative to the build's working
directory. If this value is absolute, it may be outside the build's working
directory, in which case the contents of the path may not be persisted
across build step executions, unless a `volume` for that path is specified.
- If the build specifies a `RepoSource` with `dir` and a step with a `dir`,
which specifies an absolute path, the `RepoSource` `dir` is ignored for
the step's execution.
returned: success
type: str
secretEnv:
description:
- A list of environment variables which are encrypted using a Cloud Key
Management Service crypto key. These values must be specified in the build's
`Secret`.
returned: success
type: list
timeout:
description:
- Time limit for executing this build step. If not defined, the step has
no time limit and will be allowed to continue to run until either it completes
or the build itself times out.
returned: success
type: str
timing:
description:
- Output only. Stores timing information for executing this build step.
returned: success
type: str
volumes:
description:
- List of volumes to mount into the build step.
- Each volume is created as an empty volume prior to execution of the build
step. Upon completion of the build, volumes and their contents are discarded.
- Using a named volume in only one step is not valid as it is indicative
of a build request with an incorrect configuration.
returned: success
type: complex
contains:
name:
description:
- Name of the volume to mount.
- Volume names must be unique per build step and must be valid names
for Docker volumes. Each named volume must be used by at least two
build steps.
returned: success
type: str
path:
description:
- Path at which to mount the volume.
- Paths must be absolute and cannot conflict with other volume paths
on the same build step or with certain reserved volume paths.
returned: success
type: str
waitFor:
description:
- The ID(s) of the step(s) that this build step depends on.
- This build step will not start until all the build steps in `wait_for`
have completed successfully. If `wait_for` is empty, this build step will
start when all previous build steps in the `Build.Steps` list have completed
successfully.
returned: success
type: list
artifacts:
description:
- Artifacts produced by the build that should be uploaded upon successful completion
of all build steps.
returned: success
type: complex
contains:
images:
description:
- A list of images to be pushed upon the successful completion of all build
steps.
- The images will be pushed using the builder service account's credentials.
- The digests of the pushed images will be stored in the Build resource's
results field.
- If any of the images fail to be pushed, the build is marked FAILURE.
returned: success
type: list
objects:
description:
- A list of objects to be uploaded to Cloud Storage upon successful completion
of all build steps.
- Files in the workspace matching specified paths globs will be uploaded
to the Cloud Storage location using the builder service account's credentials.
- The location and generation of the uploaded objects will be stored in
the Build resource's results field.
- If any objects fail to be pushed, the build is marked FAILURE.
returned: success
type: complex
contains:
location:
description:
- Cloud Storage bucket and optional object path, in the form "gs://bucket/path/to/somewhere/".
- Files in the workspace matching any path pattern will be uploaded
to Cloud Storage with this location as a prefix.
returned: success
type: str
paths:
description:
- Path globs used to match files in the build's workspace.
returned: success
type: list
timing:
description:
- Output only. Stores timing information for pushing all artifact objects.
returned: success
type: complex
contains:
startTime:
description:
- Start of time span.
- 'A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution
and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z"
and "2014-10-02T15:01:23.045123456Z".'
returned: success
type: str
endTime:
description:
- End of time span.
- 'A timestamp in RFC3339 UTC "Zulu" format, with nanosecond resolution
and up to nine fractional digits. Examples: "2014-10-02T15:01:23Z"
and "2014-10-02T15:01:23.045123456Z".'
returned: success
type: str
options:
description:
- Special options for this build.
returned: success
type: complex
contains:
sourceProvenanceHash:
description:
- Requested hash for SourceProvenance.
returned: success
type: list
requestedVerifyOption:
description:
- Requested verifiability options.
returned: success
type: str
machineType:
description:
- Compute Engine machine type on which to run the build.
returned: success
type: str
diskSizeGb:
description:
- Requested disk size for the VM that runs the build. Note that this is
NOT "disk free"; some of the space will be used by the operating system
and build utilities.
- Also note that this is the minimum disk size that will be allocated for
the build -- the build may run with a larger disk than requested. At present,
the maximum disk size is 1000GB; builds that request more than the maximum
are rejected with an error.
returned: success
type: int
substitutionOption:
description:
- Option to specify behavior when there is an error in the substitution
checks.
- NOTE this is always set to ALLOW_LOOSE for triggered builds and cannot
be overridden in the build configuration file.
returned: success
type: str
dynamicSubstitutions:
description:
- Option to specify whether or not to apply bash style string operations
to the substitutions.
- NOTE this is always enabled for triggered builds and cannot be overridden
in the build configuration file.
returned: success
type: bool
logStreamingOption:
description:
- Option to define build log streaming behavior to Google Cloud Storage.
returned: success
type: str
workerPool:
description:
- Option to specify a WorkerPool for the build. Format projects/{project}/workerPools/{workerPool}
This field is experimental.
returned: success
type: str
logging:
description:
- Option to specify the logging mode, which determines if and where build
logs are stored.
returned: success
type: str
env:
description:
- A list of global environment variable definitions that will exist for
all build steps in this build. If a variable is defined in both globally
and in a build step, the variable will use the build step value.
- The elements are of the form "KEY=VALUE" for the environment variable
"KEY" being given the value "VALUE".
returned: success
type: list
secretEnv:
description:
- A list of global environment variables, which are encrypted using a Cloud
Key Management Service crypto key. These values must be specified in the
build's Secret. These variables will be available to all build steps in
this build.
returned: success
type: list
volumes:
description:
- Global list of volumes to mount for ALL build steps Each volume is created
as an empty volume prior to starting the build process.
- Upon completion of the build, volumes and their contents are discarded.
Global volume names and paths cannot conflict with the volumes defined
a build step.
- Using a global volume in a build with only one step is not valid as it
is indicative of a build request with an incorrect configuration.
returned: success
type: complex
contains:
name:
description:
- Name of the volume to mount.
- Volume names must be unique per build step and must be valid names
for Docker volumes.
- Each named volume must be used by at least two build steps.
returned: success
type: str
path:
description:
- Path at which to mount the volume.
- Paths must be absolute and cannot conflict with other volume paths
on the same build step or with certain reserved volume paths.
returned: success
type: str
'''
################################################################################
# Imports
################################################################################
from ansible_collections.google.cloud.plugins.module_utils.gcp_utils import (
navigate_hash,
GcpSession,
GcpModule,
GcpRequest,
remove_nones_from_dict,
)
import json
################################################################################
# Main
################################################################################
def main():
"""Main function"""
module = GcpModule(
argument_spec=dict(
state=dict(default='present', choices=['present', 'absent'], type='str'),
id=dict(type='str'),
name=dict(type='str'),
description=dict(type='str'),
tags=dict(type='list', elements='str'),
disabled=dict(type='bool'),
substitutions=dict(type='dict'),
filename=dict(type='str'),
ignored_files=dict(type='list', elements='str'),
included_files=dict(type='list', elements='str'),
trigger_template=dict(
type='dict',
options=dict(
project_id=dict(type='str'),
repo_name=dict(default='default', type='str'),
dir=dict(type='str'),
invert_regex=dict(type='bool'),
branch_name=dict(type='str'),
tag_name=dict(type='str'),
commit_sha=dict(type='str'),
),
),
github=dict(
type='dict',
options=dict(
owner=dict(type='str'),
name=dict(type='str'),
pull_request=dict(
type='dict', options=dict(branch=dict(required=True, type='str'), comment_control=dict(type='str'), invert_regex=dict(type='bool'))
),
push=dict(type='dict', options=dict(invert_regex=dict(type='bool'), branch=dict(type='str'), tag=dict(type='str'))),
),
),
pubsub_config=dict(type='dict', options=dict(topic=dict(required=True, type='str'), service_account_email=dict(type='str'))),
webhook_config=dict(type='dict', options=dict(secret=dict(required=True, type='str', no_log=True))),
build=dict(
type='dict',
options=dict(
source=dict(
type='dict',
options=dict(
storage_source=dict(
type='dict',
options=dict(bucket=dict(required=True, type='str'), object=dict(required=True, type='str'), generation=dict(type='str')),
),
repo_source=dict(
type='dict',
options=dict(
project_id=dict(type='str'),
repo_name=dict(required=True, type='str'),
dir=dict(type='str'),
invert_regex=dict(type='bool'),
substitutions=dict(type='dict'),
branch_name=dict(type='str'),
tag_name=dict(type='str'),
commit_sha=dict(type='str'),
),
),
),
),
tags=dict(type='list', elements='str'),
images=dict(type='list', elements='str'),
substitutions=dict(type='dict'),
queue_ttl=dict(type='str'),
logs_bucket=dict(type='str'),
timeout=dict(default='600s', type='str'),
secrets=dict(
type='list',
elements='dict',
no_log=True,
options=dict(kms_key_name=dict(required=True, type='str'), secret_env=dict(type='dict', no_log=True))),
steps=dict(
required=True,
type='list',
elements='dict',
options=dict(
name=dict(required=True, type='str'),
args=dict(type='list', elements='str'),
env=dict(type='list', elements='str'),
id=dict(type='str'),
entrypoint=dict(type='str'),
dir=dict(type='str'),
secret_env=dict(type='list', elements='str', no_log=True),
timeout=dict(type='str'),
timing=dict(type='str'),
volumes=dict(
type='list', elements='dict', options=dict(name=dict(required=True, type='str'), path=dict(required=True, type='str'))
),
wait_for=dict(type='list', elements='str'),
),
),
artifacts=dict(
type='dict',
options=dict(
images=dict(type='list', elements='str'),
objects=dict(type='dict', options=dict(location=dict(type='str'), paths=dict(type='list', elements='str'))),
),
),
options=dict(
type='dict',
options=dict(
source_provenance_hash=dict(type='list', elements='str'),
requested_verify_option=dict(type='str'),
machine_type=dict(type='str'),
disk_size_gb=dict(type='int'),
substitution_option=dict(type='str'),
dynamic_substitutions=dict(type='bool'),
log_streaming_option=dict(type='str'),
worker_pool=dict(type='str'),
logging=dict(type='str'),
env=dict(type='list', elements='str'),
secret_env=dict(type='list', elements='str', no_log=True),
volumes=dict(type='list', elements='dict', options=dict(name=dict(type='str'), path=dict(type='str'))),
),
),
),
),
)
)
if not module.params['scopes']:
module.params['scopes'] = ['https://www.googleapis.com/auth/cloud-platform']
state = module.params['state']
fetch = fetch_resource(module, self_link(module))
changed = False
if fetch:
if state == 'present':
if is_different(module, fetch):
update(module, self_link(module))
fetch = fetch_resource(module, self_link(module))
changed = True
else:
delete(module, self_link(module))
fetch = {}
changed = True
else:
if state == 'present':
fetch = create(module, collection(module))
changed = True
else:
fetch = {}
fetch.update({'changed': changed})
module.exit_json(**fetch)
def create(module, link):
auth = GcpSession(module, 'cloudbuild')
return return_if_object(module, auth.post(link, resource_to_request(module)))
def update(module, link):
auth = GcpSession(module, 'cloudbuild')
return return_if_object(module, auth.patch(link, resource_to_request(module)))
def delete(module, link):
auth = GcpSession(module, 'cloudbuild')
return return_if_object(module, auth.delete(link))
def resource_to_request(module):
request = {
u'id': module.params.get('id'),
u'name': module.params.get('name'),
u'description': module.params.get('description'),
u'tags': module.params.get('tags'),
u'disabled': module.params.get('disabled'),
u'substitutions': module.params.get('substitutions'),
u'filename': module.params.get('filename'),
u'ignoredFiles': module.params.get('ignored_files'),
u'includedFiles': module.params.get('included_files'),
u'triggerTemplate': TriggerTriggertemplate(module.params.get('trigger_template', {}), module).to_request(),
u'github': TriggerGithub(module.params.get('github', {}), module).to_request(),
u'pubsubConfig': TriggerPubsubconfig(module.params.get('pubsub_config', {}), module).to_request(),
u'webhookConfig': TriggerWebhookconfig(module.params.get('webhook_config', {}), module).to_request(),
u'build': TriggerBuild(module.params.get('build', {}), module).to_request(),
}
return_vals = {}
for k, v in request.items():
if v or v is False:
return_vals[k] = v
return return_vals
def fetch_resource(module, link, allow_not_found=True):
auth = GcpSession(module, 'cloudbuild')
return return_if_object(module, auth.get(link), allow_not_found)
def self_link(module):
return "https://cloudbuild.googleapis.com/v1/projects/{project}/triggers/{id}".format(**module.params)
def collection(module):
return "https://cloudbuild.googleapis.com/v1/projects/{project}/triggers".format(**module.params)
def return_if_object(module, response, allow_not_found=False):
# If not found, return nothing.
if allow_not_found and response.status_code == 404:
return None
# If no content, return nothing.
if response.status_code == 204:
return None
try:
module.raise_for_status(response)
result = response.json()
except getattr(json.decoder, 'JSONDecodeError', ValueError):
module.fail_json(msg="Invalid JSON response with error: %s" % response.text)
if navigate_hash(result, ['error', 'errors']):
module.fail_json(msg=navigate_hash(result, ['error', 'errors']))
return result
def is_different(module, response):
request = resource_to_request(module)
response = response_to_hash(module, response)
# Remove all output-only from response.
response_vals = {}
for k, v in response.items():
if k in request:
response_vals[k] = v
request_vals = {}
for k, v in request.items():
if k in response:
request_vals[k] = v
return GcpRequest(request_vals) != GcpRequest(response_vals)
# Remove unnecessary properties from the response.
# This is for doing comparisons with Ansible's current parameters.
def response_to_hash(module, response):
return {
u'id': response.get(u'id'),
u'name': response.get(u'name'),
u'description': response.get(u'description'),
u'tags': response.get(u'tags'),
u'disabled': response.get(u'disabled'),
u'createTime': response.get(u'createTime'),
u'substitutions': response.get(u'substitutions'),
u'filename': response.get(u'filename'),
u'ignoredFiles': response.get(u'ignoredFiles'),
u'includedFiles': response.get(u'includedFiles'),
u'triggerTemplate': TriggerTriggertemplate(response.get(u'triggerTemplate', {}), module).from_response(),
u'github': TriggerGithub(response.get(u'github', {}), module).from_response(),
u'pubsubConfig': TriggerPubsubconfig(response.get(u'pubsubConfig', {}), module).from_response(),
u'webhookConfig': TriggerWebhookconfig(response.get(u'webhookConfig', {}), module).from_response(),
u'build': TriggerBuild(response.get(u'build', {}), module).from_response(),
}
class TriggerTriggertemplate(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{
u'projectId': self.request.get('project_id'),
u'repoName': self.request.get('repo_name'),
u'dir': self.request.get('dir'),
u'invertRegex': self.request.get('invert_regex'),
u'branchName': self.request.get('branch_name'),
u'tagName': self.request.get('tag_name'),
u'commitSha': self.request.get('commit_sha'),
}
)
def from_response(self):
return remove_nones_from_dict(
{
u'projectId': self.request.get(u'projectId'),
u'repoName': self.request.get(u'repoName'),
u'dir': self.request.get(u'dir'),
u'invertRegex': self.request.get(u'invertRegex'),
u'branchName': self.request.get(u'branchName'),
u'tagName': self.request.get(u'tagName'),
u'commitSha': self.request.get(u'commitSha'),
}
)
class TriggerGithub(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{
u'owner': self.request.get('owner'),
u'name': self.request.get('name'),
u'pullRequest': TriggerPullrequest(self.request.get('pull_request', {}), self.module).to_request(),
u'push': TriggerPush(self.request.get('push', {}), self.module).to_request(),
}
)
def from_response(self):
return remove_nones_from_dict(
{
u'owner': self.request.get(u'owner'),
u'name': self.request.get(u'name'),
u'pullRequest': TriggerPullrequest(self.request.get(u'pullRequest', {}), self.module).from_response(),
u'push': TriggerPush(self.request.get(u'push', {}), self.module).from_response(),
}
)
class TriggerPullrequest(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{u'branch': self.request.get('branch'), u'commentControl': self.request.get('comment_control'), u'invertRegex': self.request.get('invert_regex')}
)
def from_response(self):
return remove_nones_from_dict(
{u'branch': self.request.get(u'branch'), u'commentControl': self.request.get(u'commentControl'), u'invertRegex': self.request.get(u'invertRegex')}
)
class TriggerPush(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{u'invertRegex': self.request.get('invert_regex'), u'branch': self.request.get('branch'), u'tag': self.request.get('tag')}
)
def from_response(self):
return remove_nones_from_dict(
{u'invertRegex': self.request.get(u'invertRegex'), u'branch': self.request.get(u'branch'), u'tag': self.request.get(u'tag')}
)
class TriggerPubsubconfig(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict({u'topic': self.request.get('topic'), u'service_account_email': self.request.get('service_account_email')})
def from_response(self):
return remove_nones_from_dict({u'topic': self.request.get(u'topic'), u'service_account_email': self.request.get(u'service_account_email')})
class TriggerWebhookconfig(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict({u'secret': self.request.get('secret')})
def from_response(self):
return remove_nones_from_dict({u'secret': self.request.get(u'secret')})
class TriggerBuild(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{
u'source': TriggerSource(self.request.get('source', {}), self.module).to_request(),
u'tags': self.request.get('tags'),
u'images': self.request.get('images'),
u'substitutions': self.request.get('substitutions'),
u'queueTtl': self.request.get('queue_ttl'),
u'logsBucket': self.request.get('logs_bucket'),
u'timeout': self.request.get('timeout'),
u'secrets': TriggerSecretsArray(self.request.get('secrets', []), self.module).to_request(),
u'steps': TriggerStepsArray(self.request.get('steps', []), self.module).to_request(),
u'artifacts': TriggerArtifacts(self.request.get('artifacts', {}), self.module).to_request(),
u'options': TriggerOptions(self.request.get('options', {}), self.module).to_request(),
}
)
def from_response(self):
return remove_nones_from_dict(
{
u'source': TriggerSource(self.request.get(u'source', {}), self.module).from_response(),
u'tags': self.request.get(u'tags'),
u'images': self.request.get(u'images'),
u'substitutions': self.request.get(u'substitutions'),
u'queueTtl': self.request.get(u'queueTtl'),
u'logsBucket': self.request.get(u'logsBucket'),
u'timeout': self.request.get(u'timeout'),
u'secrets': TriggerSecretsArray(self.request.get(u'secrets', []), self.module).from_response(),
u'steps': TriggerStepsArray(self.request.get(u'steps', []), self.module).from_response(),
u'artifacts': TriggerArtifacts(self.request.get(u'artifacts', {}), self.module).from_response(),
u'options': TriggerOptions(self.request.get(u'options', {}), self.module).from_response(),
}
)
class TriggerSource(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{
u'storageSource': TriggerStoragesource(self.request.get('storage_source', {}), self.module).to_request(),
u'repoSource': TriggerReposource(self.request.get('repo_source', {}), self.module).to_request(),
}
)
def from_response(self):
return remove_nones_from_dict(
{
u'storageSource': TriggerStoragesource(self.request.get(u'storageSource', {}), self.module).from_response(),
u'repoSource': TriggerReposource(self.request.get(u'repoSource', {}), self.module).from_response(),
}
)
class TriggerStoragesource(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{u'bucket': self.request.get('bucket'), u'object': self.request.get('object'), u'generation': self.request.get('generation')}
)
def from_response(self):
return remove_nones_from_dict(
{u'bucket': self.request.get(u'bucket'), u'object': self.request.get(u'object'), u'generation': self.request.get(u'generation')}
)
class TriggerReposource(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{
u'projectId': self.request.get('project_id'),
u'repoName': self.request.get('repo_name'),
u'dir': self.request.get('dir'),
u'invertRegex': self.request.get('invert_regex'),
u'substitutions': self.request.get('substitutions'),
u'branchName': self.request.get('branch_name'),
u'tagName': self.request.get('tag_name'),
u'commitSha': self.request.get('commit_sha'),
}
)
def from_response(self):
return remove_nones_from_dict(
{
u'projectId': self.request.get(u'projectId'),
u'repoName': self.request.get(u'repoName'),
u'dir': self.request.get(u'dir'),
u'invertRegex': self.request.get(u'invertRegex'),
u'substitutions': self.request.get(u'substitutions'),
u'branchName': self.request.get(u'branchName'),
u'tagName': self.request.get(u'tagName'),
u'commitSha': self.request.get(u'commitSha'),
}
)
class TriggerSecretsArray(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = []
def to_request(self):
items = []
for item in self.request:
items.append(self._request_for_item(item))
return items
def from_response(self):
items = []
for item in self.request:
items.append(self._response_from_item(item))
return items
def _request_for_item(self, item):
return remove_nones_from_dict({u'kmsKeyName': item.get('kms_key_name'), u'secretEnv': item.get('secret_env')})
def _response_from_item(self, item):
return remove_nones_from_dict({u'kmsKeyName': item.get(u'kmsKeyName'), u'secretEnv': item.get(u'secretEnv')})
class TriggerStepsArray(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = []
def to_request(self):
items = []
for item in self.request:
items.append(self._request_for_item(item))
return items
def from_response(self):
items = []
for item in self.request:
items.append(self._response_from_item(item))
return items
def _request_for_item(self, item):
return remove_nones_from_dict(
{
u'name': item.get('name'),
u'args': item.get('args'),
u'env': item.get('env'),
u'id': item.get('id'),
u'entrypoint': item.get('entrypoint'),
u'dir': item.get('dir'),
u'secretEnv': item.get('secret_env'),
u'timeout': item.get('timeout'),
u'timing': item.get('timing'),
u'volumes': TriggerVolumesArray(item.get('volumes', []), self.module).to_request(),
u'waitFor': item.get('wait_for'),
}
)
def _response_from_item(self, item):
return remove_nones_from_dict(
{
u'name': item.get(u'name'),
u'args': item.get(u'args'),
u'env': item.get(u'env'),
u'id': item.get(u'id'),
u'entrypoint': item.get(u'entrypoint'),
u'dir': item.get(u'dir'),
u'secretEnv': item.get(u'secretEnv'),
u'timeout': item.get(u'timeout'),
u'timing': item.get(u'timing'),
u'volumes': TriggerVolumesArray(item.get(u'volumes', []), self.module).from_response(),
u'waitFor': item.get(u'waitFor'),
}
)
class TriggerVolumesArray(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = []
def to_request(self):
items = []
for item in self.request:
items.append(self._request_for_item(item))
return items
def from_response(self):
items = []
for item in self.request:
items.append(self._response_from_item(item))
return items
def _request_for_item(self, item):
return remove_nones_from_dict({u'name': item.get('name'), u'path': item.get('path')})
def _response_from_item(self, item):
return remove_nones_from_dict({u'name': item.get(u'name'), u'path': item.get(u'path')})
class TriggerArtifacts(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{u'images': self.request.get('images'), u'objects': TriggerObjects(self.request.get('objects', {}), self.module).to_request()}
)
def from_response(self):
return remove_nones_from_dict(
{u'images': self.request.get(u'images'), u'objects': TriggerObjects(self.request.get(u'objects', {}), self.module).from_response()}
)
class TriggerObjects(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict({u'location': self.request.get('location'), u'paths': self.request.get('paths')})
def from_response(self):
return remove_nones_from_dict({u'location': self.request.get(u'location'), u'paths': self.request.get(u'paths')})
class TriggerTiming(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict({u'startTime': self.request.get('start_time'), u'endTime': self.request.get('end_time')})
def from_response(self):
return remove_nones_from_dict({u'startTime': self.request.get(u'startTime'), u'endTime': self.request.get(u'endTime')})
class TriggerOptions(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = {}
def to_request(self):
return remove_nones_from_dict(
{
u'sourceProvenanceHash': self.request.get('source_provenance_hash'),
u'requestedVerifyOption': self.request.get('requested_verify_option'),
u'machineType': self.request.get('machine_type'),
u'diskSizeGb': self.request.get('disk_size_gb'),
u'substitutionOption': self.request.get('substitution_option'),
u'dynamicSubstitutions': self.request.get('dynamic_substitutions'),
u'logStreamingOption': self.request.get('log_streaming_option'),
u'workerPool': self.request.get('worker_pool'),
u'logging': self.request.get('logging'),
u'env': self.request.get('env'),
u'secretEnv': self.request.get('secret_env'),
u'volumes': TriggerVolumesArray(self.request.get('volumes', []), self.module).to_request(),
}
)
def from_response(self):
return remove_nones_from_dict(
{
u'sourceProvenanceHash': self.request.get(u'sourceProvenanceHash'),
u'requestedVerifyOption': self.request.get(u'requestedVerifyOption'),
u'machineType': self.request.get(u'machineType'),
u'diskSizeGb': self.request.get(u'diskSizeGb'),
u'substitutionOption': self.request.get(u'substitutionOption'),
u'dynamicSubstitutions': self.request.get(u'dynamicSubstitutions'),
u'logStreamingOption': self.request.get(u'logStreamingOption'),
u'workerPool': self.request.get(u'workerPool'),
u'logging': self.request.get(u'logging'),
u'env': self.request.get(u'env'),
u'secretEnv': self.request.get(u'secretEnv'),
u'volumes': TriggerVolumesArray(self.request.get(u'volumes', []), self.module).from_response(),
}
)
class TriggerVolumesArray(object):
def __init__(self, request, module):
self.module = module
if request:
self.request = request
else:
self.request = []
def to_request(self):
items = []
for item in self.request:
items.append(self._request_for_item(item))
return items
def from_response(self):
items = []
for item in self.request:
items.append(self._response_from_item(item))
return items
def _request_for_item(self, item):
return remove_nones_from_dict({u'name': item.get('name'), u'path': item.get('path')})
def _response_from_item(self, item):
return remove_nones_from_dict({u'name': item.get(u'name'), u'path': item.get(u'path')})
if __name__ == '__main__':
main()