mirror of
https://github.com/ansible-collections/community.general.git
synced 2025-04-04 17:50:32 -07:00
* Make conftest's patch_ansible_module use the context manager from .utils. * Fix test dependencies. * Use module mock utils from community.internal_test_tools. * Use DataDictLoader from community.internal_test_tools. * Use trust util from community.internal_test_tools.
262 lines
8.9 KiB
Python
262 lines
8.9 KiB
Python
# Copyright (c) Ansible project
|
|
# GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)
|
|
# SPDX-License-Identifier: GPL-3.0-or-later
|
|
|
|
|
|
from __future__ import (absolute_import, division, print_function)
|
|
__metaclass__ = type
|
|
|
|
import os
|
|
import sys
|
|
import json
|
|
|
|
import yaml
|
|
import pytest
|
|
|
|
|
|
class UTHelper(object):
|
|
TEST_SPEC_VALID_SECTIONS = ["anchors", "test_cases"]
|
|
|
|
@staticmethod
|
|
def from_spec(ansible_module, test_module, test_spec, mocks=None):
|
|
helper = UTHelper(ansible_module, test_module, test_spec=test_spec, mocks=mocks)
|
|
return helper
|
|
|
|
@staticmethod
|
|
def from_file(ansible_module, test_module, test_spec_filehandle, mocks=None):
|
|
test_spec = yaml.safe_load(test_spec_filehandle)
|
|
return UTHelper.from_spec(ansible_module, test_module, test_spec, mocks)
|
|
|
|
# @TODO: calculate the test_module_name automatically, remove one more parameter
|
|
@staticmethod
|
|
def from_module(ansible_module, test_module_name, mocks=None):
|
|
test_module = sys.modules[test_module_name]
|
|
extensions = ['.yaml', '.yml']
|
|
for ext in extensions:
|
|
test_spec_filename = test_module.__file__.replace('.py', ext)
|
|
if os.path.exists(test_spec_filename):
|
|
with open(test_spec_filename, "r") as test_spec_filehandle:
|
|
return UTHelper.from_file(ansible_module, test_module, test_spec_filehandle, mocks=mocks)
|
|
|
|
raise Exception("Cannot find test case file for {0} with one of the extensions: {1}".format(test_module.__file__, extensions))
|
|
|
|
def add_func_to_test_module(self, name, func):
|
|
setattr(self.test_module, name, func)
|
|
|
|
def __init__(self, ansible_module, test_module, test_spec, mocks=None):
|
|
self.ansible_module = ansible_module
|
|
self.test_module = test_module
|
|
self.test_cases = []
|
|
self.fixtures = {}
|
|
|
|
spec_diff = set(test_spec.keys()) - set(self.TEST_SPEC_VALID_SECTIONS)
|
|
if spec_diff:
|
|
raise ValueError("Test specification contain unknown keys: {0}".format(", ".join(spec_diff)))
|
|
|
|
self.mocks_map = {m.name: m for m in mocks} if mocks else {}
|
|
|
|
for spec_test_case in test_spec['test_cases']:
|
|
tc = ModuleTestCase.make_test_case(spec_test_case, test_module, self.mocks_map)
|
|
self.test_cases.append(tc)
|
|
self.fixtures.update(tc.fixtures)
|
|
self.set_test_func()
|
|
self.set_fixtures(self.fixtures)
|
|
|
|
@property
|
|
def runner(self):
|
|
return Runner(self.ansible_module.main)
|
|
|
|
def set_test_func(self):
|
|
@pytest.mark.parametrize('test_case', self.test_cases, ids=[tc.id for tc in self.test_cases])
|
|
@pytest.mark.usefixtures(*self.fixtures)
|
|
def _test_module(mocker, capfd, patch_ansible_module_uthelper, test_case):
|
|
"""
|
|
Run unit tests for each test case in self.test_cases
|
|
"""
|
|
args = {}
|
|
args.update(test_case.input)
|
|
if test_case.flags.get("check"):
|
|
args["_ansible_check_mode"] = test_case.flags.get("check")
|
|
if test_case.flags.get("diff"):
|
|
args["_ansible_diff"] = test_case.flags.get("diff")
|
|
with patch_ansible_module_uthelper(args):
|
|
self.runner.run(mocker, capfd, test_case)
|
|
|
|
self.add_func_to_test_module("test_module", _test_module)
|
|
|
|
return _test_module
|
|
|
|
def set_fixtures(self, fixtures):
|
|
for name, fixture in fixtures.items():
|
|
self.add_func_to_test_module(name, fixture)
|
|
|
|
|
|
class Runner:
|
|
def __init__(self, module_main):
|
|
self.module_main = module_main
|
|
self.results = None
|
|
|
|
def run(self, mocker, capfd, test_case):
|
|
test_case.setup(mocker)
|
|
self.pytest_module(capfd, test_case.flags)
|
|
test_case.check(self.results)
|
|
|
|
def pytest_module(self, capfd, flags):
|
|
if flags.get("skip"):
|
|
pytest.skip(flags.get("skip"))
|
|
if flags.get("xfail"):
|
|
pytest.xfail(flags.get("xfail"))
|
|
|
|
with pytest.raises(SystemExit):
|
|
(self.module_main)()
|
|
|
|
out, err = capfd.readouterr()
|
|
self.results = json.loads(out)
|
|
|
|
|
|
class ModuleTestCase:
|
|
def __init__(self, id, input, output, mocks, flags):
|
|
self.id = id
|
|
self.input = input
|
|
self.output = output
|
|
self.mock_specs = mocks
|
|
self.mocks = {}
|
|
self.flags = flags
|
|
|
|
self._fixtures = {}
|
|
|
|
def __str__(self):
|
|
return "<ModuleTestCase: id={id} {input}{output}mocks={mocks} flags={flags}>".format(
|
|
id=self.id,
|
|
input="input " if self.input else "",
|
|
output="output " if self.output else "",
|
|
mocks="({0})".format(", ".join(self.mocks.keys())),
|
|
flags=self.flags
|
|
)
|
|
|
|
def __repr__(self):
|
|
return "ModuleTestCase(id={id}, input={input}, output={output}, mocks={mocks}, flags={flags})".format(
|
|
id=self.id,
|
|
input=self.input,
|
|
output=self.output,
|
|
mocks=repr(self.mocks),
|
|
flags=self.flags
|
|
)
|
|
|
|
@staticmethod
|
|
def make_test_case(test_case_spec, test_module, mocks_map):
|
|
tc = ModuleTestCase(
|
|
id=test_case_spec["id"],
|
|
input=test_case_spec.get("input", {}),
|
|
output=test_case_spec.get("output", {}),
|
|
mocks=test_case_spec.get("mocks", {}),
|
|
flags=test_case_spec.get("flags", {})
|
|
)
|
|
tc.build_mocks(mocks_map)
|
|
return tc
|
|
|
|
def build_mocks(self, mocks_map):
|
|
for mock_name, mock_spec in self.mock_specs.items():
|
|
try:
|
|
mock_class = mocks_map[mock_name]
|
|
except KeyError:
|
|
raise Exception("Cannot find TestCaseMock class for: {0}".format(mock_name))
|
|
self.mocks[mock_name] = mock_class.build_mock(mock_spec)
|
|
|
|
self._fixtures.update(self.mocks[mock_name].fixtures())
|
|
|
|
@property
|
|
def fixtures(self):
|
|
return dict(self._fixtures)
|
|
|
|
def setup(self, mocker):
|
|
self.setup_testcase(mocker)
|
|
self.setup_mocks(mocker)
|
|
|
|
def check(self, results):
|
|
self.check_testcase(results)
|
|
self.check_mocks(self, results)
|
|
|
|
def setup_testcase(self, mocker):
|
|
pass
|
|
|
|
def setup_mocks(self, mocker):
|
|
for mock in self.mocks.values():
|
|
mock.setup(mocker)
|
|
|
|
def check_testcase(self, results):
|
|
print("testcase =\n%s" % repr(self))
|
|
print("results =\n%s" % results)
|
|
if 'exception' in results:
|
|
print("exception = \n%s" % results["exception"])
|
|
|
|
for test_result in self.output:
|
|
assert results[test_result] == self.output[test_result], \
|
|
"'{0}': '{1}' != '{2}'".format(test_result, results[test_result], self.output[test_result])
|
|
|
|
def check_mocks(self, test_case, results):
|
|
for mock in self.mocks.values():
|
|
mock.check(test_case, results)
|
|
|
|
|
|
class TestCaseMock:
|
|
@classmethod
|
|
def build_mock(cls, mock_specs):
|
|
return cls(mock_specs)
|
|
|
|
def __init__(self, mock_specs):
|
|
self.mock_specs = mock_specs
|
|
|
|
def fixtures(self):
|
|
return {}
|
|
|
|
def setup(self, mocker):
|
|
pass
|
|
|
|
def check(self, test_case, results):
|
|
raise NotImplementedError()
|
|
|
|
|
|
class RunCommandMock(TestCaseMock):
|
|
name = "run_command"
|
|
|
|
def __str__(self):
|
|
return "<RunCommandMock specs={specs}>".format(specs=self.mock_specs)
|
|
|
|
def __repr__(self):
|
|
return "RunCommandMock({specs})".format(specs=self.mock_specs)
|
|
|
|
def fixtures(self):
|
|
@pytest.fixture
|
|
def patch_bin(mocker):
|
|
def mockie(self_, path, *args, **kwargs):
|
|
return "/testbin/{0}".format(path)
|
|
mocker.patch('ansible.module_utils.basic.AnsibleModule.get_bin_path', mockie)
|
|
|
|
return {"patch_bin": patch_bin}
|
|
|
|
def setup(self, mocker):
|
|
def _results():
|
|
for result in [(x['rc'], x['out'], x['err']) for x in self.mock_specs]:
|
|
yield result
|
|
raise Exception("testcase has not enough run_command calls")
|
|
|
|
results = _results()
|
|
|
|
def side_effect(self_, **kwargs):
|
|
result = next(results)
|
|
if kwargs.get("check_rc", False) and result[0] != 0:
|
|
raise Exception("rc = {0}".format(result[0]))
|
|
return result
|
|
|
|
self.mock_run_cmd = mocker.patch('ansible.module_utils.basic.AnsibleModule.run_command', side_effect=side_effect)
|
|
|
|
def check(self, test_case, results):
|
|
call_args_list = [(item[0][0], item[1]) for item in self.mock_run_cmd.call_args_list]
|
|
expected_call_args_list = [(item['command'], item['environ']) for item in self.mock_specs]
|
|
print("call args list =\n%s" % call_args_list)
|
|
print("expected args list =\n%s" % expected_call_args_list)
|
|
|
|
assert self.mock_run_cmd.call_count == len(self.mock_specs), "{0} != {1}".format(self.mock_run_cmd.call_count, len(self.mock_specs))
|
|
if self.mock_run_cmd.call_count:
|
|
assert call_args_list == expected_call_args_list
|