diff options
| author | Kenny Woodson <kwoodson@redhat.com> | 2017-01-30 12:31:10 -0500 | 
|---|---|---|
| committer | GitHub <noreply@github.com> | 2017-01-30 12:31:10 -0500 | 
| commit | 8c7ccc601cab2f84a1fa3af9f4c1278dab96daf5 (patch) | |
| tree | 804d22df0bde2009452c23e40896b18afa3eb654 /roles | |
| parent | 33f6619a919b815b682bb813baf2cb4f42ea964e (diff) | |
| parent | 70849347e5cef1efe61d11eb7185d3bafd6288e9 (diff) | |
| download | openshift-8c7ccc601cab2f84a1fa3af9f4c1278dab96daf5.tar.gz openshift-8c7ccc601cab2f84a1fa3af9f4c1278dab96daf5.tar.bz2 openshift-8c7ccc601cab2f84a1fa3af9f4c1278dab96daf5.tar.xz openshift-8c7ccc601cab2f84a1fa3af9f4c1278dab96daf5.zip | |
Merge pull request #3197 from kwoodson/manage_node
Adding oadm_manage_node lib_openshift.
Diffstat (limited to 'roles')
| -rw-r--r-- | roles/lib_openshift/library/oadm_manage_node.py | 1477 | ||||
| -rw-r--r-- | roles/lib_openshift/library/oc_edit.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/library/oc_obj.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/library/oc_route.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/library/oc_scale.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/library/oc_secret.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/library/oc_version.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/src/ansible/oadm_manage_node.py | 38 | ||||
| -rw-r--r-- | roles/lib_openshift/src/class/oadm_manage_node.py | 209 | ||||
| -rw-r--r-- | roles/lib_openshift/src/doc/manage_node | 88 | ||||
| -rw-r--r-- | roles/lib_openshift/src/lib/base.py | 2 | ||||
| -rw-r--r-- | roles/lib_openshift/src/sources.yml | 9 | ||||
| -rwxr-xr-x | roles/lib_openshift/src/test/integration/oadm_manage_node.yml | 58 | ||||
| -rwxr-xr-x | roles/lib_openshift/src/test/unit/oadm_manage_node.py | 177 | 
14 files changed, 2063 insertions, 7 deletions
| diff --git a/roles/lib_openshift/library/oadm_manage_node.py b/roles/lib_openshift/library/oadm_manage_node.py new file mode 100644 index 000000000..64288f70d --- /dev/null +++ b/roles/lib_openshift/library/oadm_manage_node.py @@ -0,0 +1,1477 @@ +#!/usr/bin/env python +# pylint: disable=missing-docstring +# flake8: noqa: T001 +#     ___ ___ _  _ ___ ___    _ _____ ___ ___ +#    / __| __| \| | __| _ \  /_\_   _| __|   \ +#   | (_ | _|| .` | _||   / / _ \| | | _|| |) | +#    \___|___|_|\_|___|_|_\/_/_\_\_|_|___|___/_ _____ +#   |   \ / _ \  | \| |/ _ \_   _| | __|   \_ _|_   _| +#   | |) | (_) | | .` | (_) || |   | _|| |) | |  | | +#   |___/ \___/  |_|\_|\___/ |_|   |___|___/___| |_| +# +# Copyright 2016 Red Hat, Inc. and/or its affiliates +# and other contributors as indicated by the @author tags. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +#    http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# + +# -*- -*- -*- Begin included fragment: lib/import.py -*- -*- -*- +''' +   OpenShiftCLI class that wraps the oc commands in a subprocess +''' +# pylint: disable=too-many-lines + +from __future__ import print_function +import atexit +import json +import os +import re +import shutil +import subprocess +# pylint: disable=import-error +import ruamel.yaml as yaml +from ansible.module_utils.basic import AnsibleModule + +# -*- -*- -*- End included fragment: lib/import.py -*- -*- -*- + +# -*- -*- -*- Begin included fragment: doc/manage_node -*- -*- -*- + +DOCUMENTATION = ''' +--- +module: oadm_manage_node +short_description: Module to manage openshift nodes +description: +  - Manage openshift nodes programmatically. +options: +  kubeconfig: +    description: +    - The path for the kubeconfig file to use for authentication +    required: false +    default: /etc/origin/master/admin.kubeconfig +    aliases: [] +  debug: +    description: +    - Turn on debug output. +    required: false +    default: False +    aliases: [] +  node: +    description: +    - A list of the nodes being managed +    required: false +    default: None +    aliases: [] +  selector: +    description: +    - The selector when filtering on node labels +    required: false +    default: None +    aliases: [] +  pod_selector: +    description: +    - A selector when filtering on pod labels. +    required: false +    default: None +    aliases: [] +  evacuate: +    description: +    - Remove all pods from a node. +    required: false +    default: False +    aliases: [] +  schedulable: +    description: +    - whether or not openshift can schedule pods on this node +    required: False +    default: None +    aliases: [] +  dry_run: +    description: +    - This shows the pods that would be migrated if evacuate were called +    required: False +    default: False +    aliases: [] +  grace_period: +    description: +    - Grace period (seconds) for pods being deleted. +    required: false +    default: None +    aliases: [] +  force: +    description: +    - Whether or not to attempt to force this action in openshift +    required: false +    default: None +    aliases: [] +author: +- "Kenny Woodson <kwoodson@redhat.com>" +extends_documentation_fragment: [] +''' + +EXAMPLES = ''' +- name: oadm manage-node --schedulable=true --selector=ops_node=new +  oadm_manage_node: +    selector: ops_node=new +    schedulable: True +  register: schedout + +- name: oadm manage-node my-k8s-node-5 --evacuate +  oadm_manage_node: +    node:  my-k8s-node-5 +    evacuate: True +    force: True +''' + +# -*- -*- -*- End included fragment: doc/manage_node -*- -*- -*- + +# -*- -*- -*- Begin included fragment: ../../lib_utils/src/class/yedit.py -*- -*- -*- +# noqa: E301,E302 + + +class YeditException(Exception): +    ''' Exception class for Yedit ''' +    pass + + +# pylint: disable=too-many-public-methods +class Yedit(object): +    ''' Class to modify yaml files ''' +    re_valid_key = r"(((\[-?\d+\])|([0-9a-zA-Z%s/_-]+)).?)+$" +    re_key = r"(?:\[(-?\d+)\])|([0-9a-zA-Z%s/_-]+)" +    com_sep = set(['.', '#', '|', ':']) + +    # pylint: disable=too-many-arguments +    def __init__(self, +                 filename=None, +                 content=None, +                 content_type='yaml', +                 separator='.', +                 backup=False): +        self.content = content +        self._separator = separator +        self.filename = filename +        self.__yaml_dict = content +        self.content_type = content_type +        self.backup = backup +        self.load(content_type=self.content_type) +        if self.__yaml_dict is None: +            self.__yaml_dict = {} + +    @property +    def separator(self): +        ''' getter method for yaml_dict ''' +        return self._separator + +    @separator.setter +    def separator(self): +        ''' getter method for yaml_dict ''' +        return self._separator + +    @property +    def yaml_dict(self): +        ''' getter method for yaml_dict ''' +        return self.__yaml_dict + +    @yaml_dict.setter +    def yaml_dict(self, value): +        ''' setter method for yaml_dict ''' +        self.__yaml_dict = value + +    @staticmethod +    def parse_key(key, sep='.'): +        '''parse the key allowing the appropriate separator''' +        common_separators = list(Yedit.com_sep - set([sep])) +        return re.findall(Yedit.re_key % ''.join(common_separators), key) + +    @staticmethod +    def valid_key(key, sep='.'): +        '''validate the incoming key''' +        common_separators = list(Yedit.com_sep - set([sep])) +        if not re.match(Yedit.re_valid_key % ''.join(common_separators), key): +            return False + +        return True + +    @staticmethod +    def remove_entry(data, key, sep='.'): +        ''' remove data at location key ''' +        if key == '' and isinstance(data, dict): +            data.clear() +            return True +        elif key == '' and isinstance(data, list): +            del data[:] +            return True + +        if not (key and Yedit.valid_key(key, sep)) and \ +           isinstance(data, (list, dict)): +            return None + +        key_indexes = Yedit.parse_key(key, sep) +        for arr_ind, dict_key in key_indexes[:-1]: +            if dict_key and isinstance(data, dict): +                data = data.get(dict_key, None) +            elif (arr_ind and isinstance(data, list) and +                  int(arr_ind) <= len(data) - 1): +                data = data[int(arr_ind)] +            else: +                return None + +        # process last index for remove +        # expected list entry +        if key_indexes[-1][0]: +            if isinstance(data, list) and int(key_indexes[-1][0]) <= len(data) - 1:  # noqa: E501 +                del data[int(key_indexes[-1][0])] +                return True + +        # expected dict entry +        elif key_indexes[-1][1]: +            if isinstance(data, dict): +                del data[key_indexes[-1][1]] +                return True + +    @staticmethod +    def add_entry(data, key, item=None, sep='.'): +        ''' Get an item from a dictionary with key notation a.b.c +            d = {'a': {'b': 'c'}}} +            key = a#b +            return c +        ''' +        if key == '': +            pass +        elif (not (key and Yedit.valid_key(key, sep)) and +              isinstance(data, (list, dict))): +            return None + +        key_indexes = Yedit.parse_key(key, sep) +        for arr_ind, dict_key in key_indexes[:-1]: +            if dict_key: +                if isinstance(data, dict) and dict_key in data and data[dict_key]:  # noqa: E501 +                    data = data[dict_key] +                    continue + +                elif data and not isinstance(data, dict): +                    return None + +                data[dict_key] = {} +                data = data[dict_key] + +            elif (arr_ind and isinstance(data, list) and +                  int(arr_ind) <= len(data) - 1): +                data = data[int(arr_ind)] +            else: +                return None + +        if key == '': +            data = item + +        # process last index for add +        # expected list entry +        elif key_indexes[-1][0] and isinstance(data, list) and int(key_indexes[-1][0]) <= len(data) - 1:  # noqa: E501 +            data[int(key_indexes[-1][0])] = item + +        # expected dict entry +        elif key_indexes[-1][1] and isinstance(data, dict): +            data[key_indexes[-1][1]] = item + +        return data + +    @staticmethod +    def get_entry(data, key, sep='.'): +        ''' Get an item from a dictionary with key notation a.b.c +            d = {'a': {'b': 'c'}}} +            key = a.b +            return c +        ''' +        if key == '': +            pass +        elif (not (key and Yedit.valid_key(key, sep)) and +              isinstance(data, (list, dict))): +            return None + +        key_indexes = Yedit.parse_key(key, sep) +        for arr_ind, dict_key in key_indexes: +            if dict_key and isinstance(data, dict): +                data = data.get(dict_key, None) +            elif (arr_ind and isinstance(data, list) and +                  int(arr_ind) <= len(data) - 1): +                data = data[int(arr_ind)] +            else: +                return None + +        return data + +    def write(self): +        ''' write to file ''' +        if not self.filename: +            raise YeditException('Please specify a filename.') + +        if self.backup and self.file_exists(): +            shutil.copy(self.filename, self.filename + '.orig') + +        tmp_filename = self.filename + '.yedit' +        with open(tmp_filename, 'w') as yfd: +            # pylint: disable=no-member +            if hasattr(self.yaml_dict, 'fa'): +                self.yaml_dict.fa.set_block_style() + +            yfd.write(yaml.dump(self.yaml_dict, Dumper=yaml.RoundTripDumper)) + +        os.rename(tmp_filename, self.filename) + +        return (True, self.yaml_dict) + +    def read(self): +        ''' read from file ''' +        # check if it exists +        if self.filename is None or not self.file_exists(): +            return None + +        contents = None +        with open(self.filename) as yfd: +            contents = yfd.read() + +        return contents + +    def file_exists(self): +        ''' return whether file exists ''' +        if os.path.exists(self.filename): +            return True + +        return False + +    def load(self, content_type='yaml'): +        ''' return yaml file ''' +        contents = self.read() + +        if not contents and not self.content: +            return None + +        if self.content: +            if isinstance(self.content, dict): +                self.yaml_dict = self.content +                return self.yaml_dict +            elif isinstance(self.content, str): +                contents = self.content + +        # check if it is yaml +        try: +            if content_type == 'yaml' and contents: +                self.yaml_dict = yaml.load(contents, yaml.RoundTripLoader) +                # pylint: disable=no-member +                if hasattr(self.yaml_dict, 'fa'): +                    self.yaml_dict.fa.set_block_style() +            elif content_type == 'json' and contents: +                self.yaml_dict = json.loads(contents) +        except yaml.YAMLError as err: +            # Error loading yaml or json +            raise YeditException('Problem with loading yaml file. %s' % err) + +        return self.yaml_dict + +    def get(self, key): +        ''' get a specified key''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, key, self.separator) +        except KeyError: +            entry = None + +        return entry + +    def pop(self, path, key_or_item): +        ''' remove a key, value pair from a dict or an item for a list''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        except KeyError: +            entry = None + +        if entry is None: +            return (False, self.yaml_dict) + +        if isinstance(entry, dict): +            # pylint: disable=no-member,maybe-no-member +            if key_or_item in entry: +                entry.pop(key_or_item) +                return (True, self.yaml_dict) +            return (False, self.yaml_dict) + +        elif isinstance(entry, list): +            # pylint: disable=no-member,maybe-no-member +            ind = None +            try: +                ind = entry.index(key_or_item) +            except ValueError: +                return (False, self.yaml_dict) + +            entry.pop(ind) +            return (True, self.yaml_dict) + +        return (False, self.yaml_dict) + +    def delete(self, path): +        ''' remove path from a dict''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        except KeyError: +            entry = None + +        if entry is None: +            return (False, self.yaml_dict) + +        result = Yedit.remove_entry(self.yaml_dict, path, self.separator) +        if not result: +            return (False, self.yaml_dict) + +        return (True, self.yaml_dict) + +    def exists(self, path, value): +        ''' check if value exists at path''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        except KeyError: +            entry = None + +        if isinstance(entry, list): +            if value in entry: +                return True +            return False + +        elif isinstance(entry, dict): +            if isinstance(value, dict): +                rval = False +                for key, val in value.items(): +                    if entry[key] != val: +                        rval = False +                        break +                else: +                    rval = True +                return rval + +            return value in entry + +        return entry == value + +    def append(self, path, value): +        '''append value to a list''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        except KeyError: +            entry = None + +        if entry is None: +            self.put(path, []) +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        if not isinstance(entry, list): +            return (False, self.yaml_dict) + +        # pylint: disable=no-member,maybe-no-member +        entry.append(value) +        return (True, self.yaml_dict) + +    # pylint: disable=too-many-arguments +    def update(self, path, value, index=None, curr_value=None): +        ''' put path, value into a dict ''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        except KeyError: +            entry = None + +        if isinstance(entry, dict): +            # pylint: disable=no-member,maybe-no-member +            if not isinstance(value, dict): +                raise YeditException('Cannot replace key, value entry in ' + +                                     'dict with non-dict type. value=[%s] [%s]' % (value, type(value)))  # noqa: E501 + +            entry.update(value) +            return (True, self.yaml_dict) + +        elif isinstance(entry, list): +            # pylint: disable=no-member,maybe-no-member +            ind = None +            if curr_value: +                try: +                    ind = entry.index(curr_value) +                except ValueError: +                    return (False, self.yaml_dict) + +            elif index is not None: +                ind = index + +            if ind is not None and entry[ind] != value: +                entry[ind] = value +                return (True, self.yaml_dict) + +            # see if it exists in the list +            try: +                ind = entry.index(value) +            except ValueError: +                # doesn't exist, append it +                entry.append(value) +                return (True, self.yaml_dict) + +            # already exists, return +            if ind is not None: +                return (False, self.yaml_dict) +        return (False, self.yaml_dict) + +    def put(self, path, value): +        ''' put path, value into a dict ''' +        try: +            entry = Yedit.get_entry(self.yaml_dict, path, self.separator) +        except KeyError: +            entry = None + +        if entry == value: +            return (False, self.yaml_dict) + +        # deepcopy didn't work +        tmp_copy = yaml.load(yaml.round_trip_dump(self.yaml_dict, +                                                  default_flow_style=False), +                             yaml.RoundTripLoader) +        # pylint: disable=no-member +        if hasattr(self.yaml_dict, 'fa'): +            tmp_copy.fa.set_block_style() +        result = Yedit.add_entry(tmp_copy, path, value, self.separator) +        if not result: +            return (False, self.yaml_dict) + +        self.yaml_dict = tmp_copy + +        return (True, self.yaml_dict) + +    def create(self, path, value): +        ''' create a yaml file ''' +        if not self.file_exists(): +            # deepcopy didn't work +            tmp_copy = yaml.load(yaml.round_trip_dump(self.yaml_dict, default_flow_style=False),  # noqa: E501 +                                 yaml.RoundTripLoader) +            # pylint: disable=no-member +            if hasattr(self.yaml_dict, 'fa'): +                tmp_copy.fa.set_block_style() +            result = Yedit.add_entry(tmp_copy, path, value, self.separator) +            if result: +                self.yaml_dict = tmp_copy +                return (True, self.yaml_dict) + +        return (False, self.yaml_dict) + +    @staticmethod +    def get_curr_value(invalue, val_type): +        '''return the current value''' +        if invalue is None: +            return None + +        curr_value = invalue +        if val_type == 'yaml': +            curr_value = yaml.load(invalue) +        elif val_type == 'json': +            curr_value = json.loads(invalue) + +        return curr_value + +    @staticmethod +    def parse_value(inc_value, vtype=''): +        '''determine value type passed''' +        true_bools = ['y', 'Y', 'yes', 'Yes', 'YES', 'true', 'True', 'TRUE', +                      'on', 'On', 'ON', ] +        false_bools = ['n', 'N', 'no', 'No', 'NO', 'false', 'False', 'FALSE', +                       'off', 'Off', 'OFF'] + +        # It came in as a string but you didn't specify value_type as string +        # we will convert to bool if it matches any of the above cases +        if isinstance(inc_value, str) and 'bool' in vtype: +            if inc_value not in true_bools and inc_value not in false_bools: +                raise YeditException('Not a boolean type. str=[%s] vtype=[%s]' +                                     % (inc_value, vtype)) +        elif isinstance(inc_value, bool) and 'str' in vtype: +            inc_value = str(inc_value) + +        # If vtype is not str then go ahead and attempt to yaml load it. +        if isinstance(inc_value, str) and 'str' not in vtype: +            try: +                inc_value = yaml.load(inc_value) +            except Exception: +                raise YeditException('Could not determine type of incoming ' + +                                     'value. value=[%s] vtype=[%s]' +                                     % (type(inc_value), vtype)) + +        return inc_value + +    # pylint: disable=too-many-return-statements,too-many-branches +    @staticmethod +    def run_ansible(module): +        '''perform the idempotent crud operations''' +        yamlfile = Yedit(filename=module.params['src'], +                         backup=module.params['backup'], +                         separator=module.params['separator']) + +        if module.params['src']: +            rval = yamlfile.load() + +            if yamlfile.yaml_dict is None and \ +               module.params['state'] != 'present': +                return {'failed': True, +                        'msg': 'Error opening file [%s].  Verify that the ' + +                               'file exists, that it is has correct' + +                               ' permissions, and is valid yaml.'} + +        if module.params['state'] == 'list': +            if module.params['content']: +                content = Yedit.parse_value(module.params['content'], +                                            module.params['content_type']) +                yamlfile.yaml_dict = content + +            if module.params['key']: +                rval = yamlfile.get(module.params['key']) or {} + +            return {'changed': False, 'result': rval, 'state': "list"} + +        elif module.params['state'] == 'absent': +            if module.params['content']: +                content = Yedit.parse_value(module.params['content'], +                                            module.params['content_type']) +                yamlfile.yaml_dict = content + +            if module.params['update']: +                rval = yamlfile.pop(module.params['key'], +                                    module.params['value']) +            else: +                rval = yamlfile.delete(module.params['key']) + +            if rval[0] and module.params['src']: +                yamlfile.write() + +            return {'changed': rval[0], 'result': rval[1], 'state': "absent"} + +        elif module.params['state'] == 'present': +            # check if content is different than what is in the file +            if module.params['content']: +                content = Yedit.parse_value(module.params['content'], +                                            module.params['content_type']) + +                # We had no edits to make and the contents are the same +                if yamlfile.yaml_dict == content and \ +                   module.params['value'] is None: +                    return {'changed': False, +                            'result': yamlfile.yaml_dict, +                            'state': "present"} + +                yamlfile.yaml_dict = content + +            # we were passed a value; parse it +            if module.params['value']: +                value = Yedit.parse_value(module.params['value'], +                                          module.params['value_type']) +                key = module.params['key'] +                if module.params['update']: +                    # pylint: disable=line-too-long +                    curr_value = Yedit.get_curr_value(Yedit.parse_value(module.params['curr_value']),  # noqa: E501 +                                                      module.params['curr_value_format'])  # noqa: E501 + +                    rval = yamlfile.update(key, value, module.params['index'], curr_value)  # noqa: E501 + +                elif module.params['append']: +                    rval = yamlfile.append(key, value) +                else: +                    rval = yamlfile.put(key, value) + +                if rval[0] and module.params['src']: +                    yamlfile.write() + +                return {'changed': rval[0], +                        'result': rval[1], 'state': "present"} + +            # no edits to make +            if module.params['src']: +                # pylint: disable=redefined-variable-type +                rval = yamlfile.write() +                return {'changed': rval[0], +                        'result': rval[1], +                        'state': "present"} + +        return {'failed': True, 'msg': 'Unkown state passed'} + +# -*- -*- -*- End included fragment: ../../lib_utils/src/class/yedit.py -*- -*- -*- + +# -*- -*- -*- Begin included fragment: lib/base.py -*- -*- -*- +# pylint: disable=too-many-lines +# noqa: E301,E302,E303,T001 + + +class OpenShiftCLIError(Exception): +    '''Exception class for openshiftcli''' +    pass + + +# pylint: disable=too-few-public-methods +class OpenShiftCLI(object): +    ''' Class to wrap the command line tools ''' +    def __init__(self, +                 namespace, +                 kubeconfig='/etc/origin/master/admin.kubeconfig', +                 verbose=False, +                 all_namespaces=False): +        ''' Constructor for OpenshiftCLI ''' +        self.namespace = namespace +        self.verbose = verbose +        self.kubeconfig = kubeconfig +        self.all_namespaces = all_namespaces + +    # Pylint allows only 5 arguments to be passed. +    # pylint: disable=too-many-arguments +    def _replace_content(self, resource, rname, content, force=False, sep='.'): +        ''' replace the current object with the content ''' +        res = self._get(resource, rname) +        if not res['results']: +            return res + +        fname = '/tmp/%s' % rname +        yed = Yedit(fname, res['results'][0], separator=sep) +        changes = [] +        for key, value in content.items(): +            changes.append(yed.put(key, value)) + +        if any([change[0] for change in changes]): +            yed.write() + +            atexit.register(Utils.cleanup, [fname]) + +            return self._replace(fname, force) + +        return {'returncode': 0, 'updated': False} + +    def _replace(self, fname, force=False): +        '''replace the current object with oc replace''' +        cmd = ['replace', '-f', fname] +        if force: +            cmd.append('--force') +        return self.openshift_cmd(cmd) + +    def _create_from_content(self, rname, content): +        '''create a temporary file and then call oc create on it''' +        fname = '/tmp/%s' % rname +        yed = Yedit(fname, content=content) +        yed.write() + +        atexit.register(Utils.cleanup, [fname]) + +        return self._create(fname) + +    def _create(self, fname): +        '''call oc create on a filename''' +        return self.openshift_cmd(['create', '-f', fname]) + +    def _delete(self, resource, rname, selector=None): +        '''call oc delete on a resource''' +        cmd = ['delete', resource, rname] +        if selector: +            cmd.append('--selector=%s' % selector) + +        return self.openshift_cmd(cmd) + +    def _process(self, template_name, create=False, params=None, template_data=None):  # noqa: E501 +        '''process a template + +           template_name: the name of the template to process +           create: whether to send to oc create after processing +           params: the parameters for the template +           template_data: the incoming template's data; instead of a file +        ''' +        cmd = ['process'] +        if template_data: +            cmd.extend(['-f', '-']) +        else: +            cmd.append(template_name) +        if params: +            param_str = ["%s=%s" % (key, value) for key, value in params.items()] +            cmd.append('-v') +            cmd.extend(param_str) + +        results = self.openshift_cmd(cmd, output=True, input_data=template_data) + +        if results['returncode'] != 0 or not create: +            return results + +        fname = '/tmp/%s' % template_name +        yed = Yedit(fname, results['results']) +        yed.write() + +        atexit.register(Utils.cleanup, [fname]) + +        return self.openshift_cmd(['create', '-f', fname]) + +    def _get(self, resource, rname=None, selector=None): +        '''return a resource by name ''' +        cmd = ['get', resource] +        if selector: +            cmd.append('--selector=%s' % selector) +        elif rname: +            cmd.append(rname) + +        cmd.extend(['-o', 'json']) + +        rval = self.openshift_cmd(cmd, output=True) + +        # Ensure results are retuned in an array +        if 'items' in rval: +            rval['results'] = rval['items'] +        elif not isinstance(rval['results'], list): +            rval['results'] = [rval['results']] + +        return rval + +    def _schedulable(self, node=None, selector=None, schedulable=True): +        ''' perform oadm manage-node scheduable ''' +        cmd = ['manage-node'] +        if node: +            cmd.extend(node) +        else: +            cmd.append('--selector=%s' % selector) + +        cmd.append('--schedulable=%s' % schedulable) + +        return self.openshift_cmd(cmd, oadm=True, output=True, output_type='raw')  # noqa: E501 + +    def _list_pods(self, node=None, selector=None, pod_selector=None): +        ''' perform oadm list pods + +            node: the node in which to list pods +            selector: the label selector filter if provided +            pod_selector: the pod selector filter if provided +        ''' +        cmd = ['manage-node'] +        if node: +            cmd.extend(node) +        else: +            cmd.append('--selector=%s' % selector) + +        if pod_selector: +            cmd.append('--pod-selector=%s' % pod_selector) + +        cmd.extend(['--list-pods', '-o', 'json']) + +        return self.openshift_cmd(cmd, oadm=True, output=True, output_type='raw') + +    # pylint: disable=too-many-arguments +    def _evacuate(self, node=None, selector=None, pod_selector=None, dry_run=False, grace_period=None, force=False): +        ''' perform oadm manage-node evacuate ''' +        cmd = ['manage-node'] +        if node: +            cmd.extend(node) +        else: +            cmd.append('--selector=%s' % selector) + +        if dry_run: +            cmd.append('--dry-run') + +        if pod_selector: +            cmd.append('--pod-selector=%s' % pod_selector) + +        if grace_period: +            cmd.append('--grace-period=%s' % int(grace_period)) + +        if force: +            cmd.append('--force') + +        cmd.append('--evacuate') + +        return self.openshift_cmd(cmd, oadm=True, output=True, output_type='raw') + +    def _version(self): +        ''' return the openshift version''' +        return self.openshift_cmd(['version'], output=True, output_type='raw') + +    def _import_image(self, url=None, name=None, tag=None): +        ''' perform image import ''' +        cmd = ['import-image'] + +        image = '{0}'.format(name) +        if tag: +            image += ':{0}'.format(tag) + +        cmd.append(image) + +        if url: +            cmd.append('--from={0}/{1}'.format(url, image)) + +        cmd.append('-n{0}'.format(self.namespace)) + +        cmd.append('--confirm') +        return self.openshift_cmd(cmd) + +    # pylint: disable=too-many-arguments,too-many-branches +    def openshift_cmd(self, cmd, oadm=False, output=False, output_type='json', input_data=None): +        '''Base command for oc ''' +        cmds = [] +        if oadm: +            cmds = ['/usr/bin/oadm'] +        else: +            cmds = ['/usr/bin/oc'] + +        if self.all_namespaces: +            cmds.extend(['--all-namespaces']) +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501 +            cmds.extend(['-n', self.namespace]) + +        cmds.extend(cmd) + +        rval = {} +        results = '' +        err = None + +        if self.verbose: +            print(' '.join(cmds)) + +        proc = subprocess.Popen(cmds, +                                stdin=subprocess.PIPE, +                                stdout=subprocess.PIPE, +                                stderr=subprocess.PIPE, +                                env={'KUBECONFIG': self.kubeconfig}) + +        stdout, stderr = proc.communicate(input_data) +        rval = {"returncode": proc.returncode, +                "results": results, +                "cmd": ' '.join(cmds)} + +        if proc.returncode == 0: +            if output: +                if output_type == 'json': +                    try: +                        rval['results'] = json.loads(stdout) +                    except ValueError as err: +                        if "No JSON object could be decoded" in err.args: +                            err = err.args +                elif output_type == 'raw': +                    rval['results'] = stdout + +            if self.verbose: +                print("STDOUT: {0}".format(stdout)) +                print("STDERR: {0}".format(stderr)) + +            if err: +                rval.update({"err": err, +                             "stderr": stderr, +                             "stdout": stdout, +                             "cmd": cmds}) + +        else: +            rval.update({"stderr": stderr, +                         "stdout": stdout, +                         "results": {}}) + +        return rval + + +class Utils(object): +    ''' utilities for openshiftcli modules ''' +    @staticmethod +    def create_file(rname, data, ftype='yaml'): +        ''' create a file in tmp with name and contents''' +        path = os.path.join('/tmp', rname) +        with open(path, 'w') as fds: +            if ftype == 'yaml': +                fds.write(yaml.dump(data, Dumper=yaml.RoundTripDumper)) + +            elif ftype == 'json': +                fds.write(json.dumps(data)) +            else: +                fds.write(data) + +        # Register cleanup when module is done +        atexit.register(Utils.cleanup, [path]) +        return path + +    @staticmethod +    def create_files_from_contents(content, content_type=None): +        '''Turn an array of dict: filename, content into a files array''' +        if not isinstance(content, list): +            content = [content] +        files = [] +        for item in content: +            path = Utils.create_file(item['path'], item['data'], ftype=content_type) +            files.append({'name': os.path.basename(path), 'path': path}) +        return files + +    @staticmethod +    def cleanup(files): +        '''Clean up on exit ''' +        for sfile in files: +            if os.path.exists(sfile): +                if os.path.isdir(sfile): +                    shutil.rmtree(sfile) +                elif os.path.isfile(sfile): +                    os.remove(sfile) + +    @staticmethod +    def exists(results, _name): +        ''' Check to see if the results include the name ''' +        if not results: +            return False + +        if Utils.find_result(results, _name): +            return True + +        return False + +    @staticmethod +    def find_result(results, _name): +        ''' Find the specified result by name''' +        rval = None +        for result in results: +            if 'metadata' in result and result['metadata']['name'] == _name: +                rval = result +                break + +        return rval + +    @staticmethod +    def get_resource_file(sfile, sfile_type='yaml'): +        ''' return the service file ''' +        contents = None +        with open(sfile) as sfd: +            contents = sfd.read() + +        if sfile_type == 'yaml': +            contents = yaml.load(contents, yaml.RoundTripLoader) +        elif sfile_type == 'json': +            contents = json.loads(contents) + +        return contents + +    @staticmethod +    def filter_versions(stdout): +        ''' filter the oc version output ''' + +        version_dict = {} +        version_search = ['oc', 'openshift', 'kubernetes'] + +        for line in stdout.strip().split('\n'): +            for term in version_search: +                if not line: +                    continue +                if line.startswith(term): +                    version_dict[term] = line.split()[-1] + +        # horrible hack to get openshift version in Openshift 3.2 +        #  By default "oc version in 3.2 does not return an "openshift" version +        if "openshift" not in version_dict: +            version_dict["openshift"] = version_dict["oc"] + +        return version_dict + +    @staticmethod +    def add_custom_versions(versions): +        ''' create custom versions strings ''' + +        versions_dict = {} + +        for tech, version in versions.items(): +            # clean up "-" from version +            if "-" in version: +                version = version.split("-")[0] + +            if version.startswith('v'): +                versions_dict[tech + '_numeric'] = version[1:].split('+')[0] +                # "v3.3.0.33" is what we have, we want "3.3" +                versions_dict[tech + '_short'] = version[1:4] + +        return versions_dict + +    @staticmethod +    def openshift_installed(): +        ''' check if openshift is installed ''' +        import yum + +        yum_base = yum.YumBase() +        if yum_base.rpmdb.searchNevra(name='atomic-openshift'): +            return True + +        return False + +    # Disabling too-many-branches.  This is a yaml dictionary comparison function +    # pylint: disable=too-many-branches,too-many-return-statements,too-many-statements +    @staticmethod +    def check_def_equal(user_def, result_def, skip_keys=None, debug=False): +        ''' Given a user defined definition, compare it with the results given back by our query.  ''' + +        # Currently these values are autogenerated and we do not need to check them +        skip = ['metadata', 'status'] +        if skip_keys: +            skip.extend(skip_keys) + +        for key, value in result_def.items(): +            if key in skip: +                continue + +            # Both are lists +            if isinstance(value, list): +                if key not in user_def: +                    if debug: +                        print('User data does not have key [%s]' % key) +                        print('User data: %s' % user_def) +                    return False + +                if not isinstance(user_def[key], list): +                    if debug: +                        print('user_def[key] is not a list key=[%s] user_def[key]=%s' % (key, user_def[key])) +                    return False + +                if len(user_def[key]) != len(value): +                    if debug: +                        print("List lengths are not equal.") +                        print("key=[%s]: user_def[%s] != value[%s]" % (key, len(user_def[key]), len(value))) +                        print("user_def: %s" % user_def[key]) +                        print("value: %s" % value) +                    return False + +                for values in zip(user_def[key], value): +                    if isinstance(values[0], dict) and isinstance(values[1], dict): +                        if debug: +                            print('sending list - list') +                            print(type(values[0])) +                            print(type(values[1])) +                        result = Utils.check_def_equal(values[0], values[1], skip_keys=skip_keys, debug=debug) +                        if not result: +                            print('list compare returned false') +                            return False + +                    elif value != user_def[key]: +                        if debug: +                            print('value should be identical') +                            print(value) +                            print(user_def[key]) +                        return False + +            # recurse on a dictionary +            elif isinstance(value, dict): +                if key not in user_def: +                    if debug: +                        print("user_def does not have key [%s]" % key) +                    return False +                if not isinstance(user_def[key], dict): +                    if debug: +                        print("dict returned false: not instance of dict") +                    return False + +                # before passing ensure keys match +                api_values = set(value.keys()) - set(skip) +                user_values = set(user_def[key].keys()) - set(skip) +                if api_values != user_values: +                    if debug: +                        print("keys are not equal in dict") +                        print(api_values) +                        print(user_values) +                    return False + +                result = Utils.check_def_equal(user_def[key], value, skip_keys=skip_keys, debug=debug) +                if not result: +                    if debug: +                        print("dict returned false") +                        print(result) +                    return False + +            # Verify each key, value pair is the same +            else: +                if key not in user_def or value != user_def[key]: +                    if debug: +                        print("value not equal; user_def does not have key") +                        print(key) +                        print(value) +                        if key in user_def: +                            print(user_def[key]) +                    return False + +        if debug: +            print('returning true') +        return True + + +class OpenShiftCLIConfig(object): +    '''Generic Config''' +    def __init__(self, rname, namespace, kubeconfig, options): +        self.kubeconfig = kubeconfig +        self.name = rname +        self.namespace = namespace +        self._options = options + +    @property +    def config_options(self): +        ''' return config options ''' +        return self._options + +    def to_option_list(self): +        '''return all options as a string''' +        return self.stringify() + +    def stringify(self): +        ''' return the options hash as cli params in a string ''' +        rval = [] +        for key, data in self.config_options.items(): +            if data['include'] \ +               and (data['value'] or isinstance(data['value'], int)): +                rval.append('--%s=%s' % (key.replace('_', '-'), data['value'])) + +        return rval + + +# -*- -*- -*- End included fragment: lib/base.py -*- -*- -*- + +# -*- -*- -*- Begin included fragment: class/oadm_manage_node.py -*- -*- -*- + + +class ManageNodeException(Exception): +    ''' manage-node exception class ''' +    pass + + +class ManageNodeConfig(OpenShiftCLIConfig): +    ''' ManageNodeConfig is a DTO for the manage-node command.''' +    def __init__(self, kubeconfig, node_options): +        super(ManageNodeConfig, self).__init__(None, None, kubeconfig, node_options) + + +# pylint: disable=too-many-instance-attributes +class ManageNode(OpenShiftCLI): +    ''' Class to wrap the oc command line tools ''' + +    # pylint allows 5 +    # pylint: disable=too-many-arguments +    def __init__(self, +                 config, +                 verbose=False): +        ''' Constructor for ManageNode ''' +        super(ManageNode, self).__init__(None, config.kubeconfig) +        self.config = config + +    def evacuate(self): +        ''' formulate the params and run oadm manage-node ''' +        return self._evacuate(node=self.config.config_options['node']['value'], +                              selector=self.config.config_options['selector']['value'], +                              pod_selector=self.config.config_options['pod_selector']['value'], +                              dry_run=self.config.config_options['dry_run']['value'], +                              grace_period=self.config.config_options['grace_period']['value'], +                              force=self.config.config_options['force']['value'], +                             ) +    def get_nodes(self, node=None, selector=''): +        '''perform oc get node''' +        _node = None +        _sel = None +        if node: +            _node = node +        if selector: +            _sel = selector + +        results = self._get('node', rname=_node, selector=_sel) +        if results['returncode'] != 0: +            return results + +        nodes = [] +        items = None +        if results['results'][0]['kind'] == 'List': +            items = results['results'][0]['items'] +        else: +            items = results['results'] + +        for node in items: +            _node = {} +            _node['name'] = node['metadata']['name'] +            _node['schedulable'] = True +            if 'unschedulable' in node['spec']: +                _node['schedulable'] = False +            nodes.append(_node) + +        return nodes + +    def get_pods_from_node(self, node, pod_selector=None): +        '''return pods for a node''' +        results = self._list_pods(node=[node], pod_selector=pod_selector) + +        if results['returncode'] != 0: +            return results + +        # When a selector or node is matched it is returned along with the json. +        # We are going to split the results based on the regexp and then +        # load the json for each matching node. +        # Before we return we are going to loop over the results and pull out the node names. +        # {'node': [pod, pod], 'node': [pod, pod]} +        # 3.2 includes the following lines in stdout: "Listing matched pods on node:" +        all_pods = [] +        if "Listing matched" in results['results']: +            listing_match = re.compile('\n^Listing matched.*$\n', flags=re.MULTILINE) +            pods = listing_match.split(results['results']) +            for pod in pods: +                if pod: +                    all_pods.extend(json.loads(pod)['items']) + +        # 3.3 specific +        else: +            # this is gross but I filed a bug... +            # https://bugzilla.redhat.com/show_bug.cgi?id=1381621 +            # build our own json from the output. +            all_pods = json.loads(results['results'])['items'] + +        return all_pods + +    def list_pods(self): +        ''' run oadm manage-node --list-pods''' +        _nodes = self.config.config_options['node']['value'] +        _selector = self.config.config_options['selector']['value'] +        _pod_selector = self.config.config_options['pod_selector']['value'] + +        if not _nodes: +            _nodes = self.get_nodes(selector=_selector) +        else: +            _nodes = [{'name': name} for name in _nodes] + +        all_pods = {} +        for node in _nodes: +            results = self.get_pods_from_node(node['name'], pod_selector=_pod_selector) +            if isinstance(results, dict): +                return results +            all_pods[node['name']] = results + +        results = {} +        results['nodes'] = all_pods +        results['returncode'] = 0 +        return results + +    def schedulable(self): +        '''oadm manage-node call for making nodes unschedulable''' +        nodes = self.config.config_options['node']['value'] +        selector = self.config.config_options['selector']['value'] + +        if not nodes: +            nodes = self.get_nodes(selector=selector) +        else: +            tmp_nodes = [] +            for name in nodes: +                tmp_result = self.get_nodes(name) +                if isinstance(tmp_result, dict): +                    tmp_nodes.append(tmp_result) +                    continue +                tmp_nodes.extend(tmp_result) +            nodes = tmp_nodes + +        # This is a short circuit based on the way we fetch nodes. +        # If node is a dict/list then we've already fetched them. +        for node in nodes: +            if isinstance(node, dict) and 'returncode' in node: +                return {'results': nodes, 'returncode': node['returncode']} +            if isinstance(node, list) and 'returncode' in node[0]: +                return {'results': nodes, 'returncode': node[0]['returncode']} +        # check all the nodes that were returned and verify they are: +        # node['schedulable'] == self.config.config_options['schedulable']['value'] +        if any([node['schedulable'] != self.config.config_options['schedulable']['value'] for node in nodes]): + +            results = self._schedulable(node=self.config.config_options['node']['value'], +                                        selector=self.config.config_options['selector']['value'], +                                        schedulable=self.config.config_options['schedulable']['value']) + +            # 'NAME                            STATUS    AGE\\nip-172-31-49-140.ec2.internal   Ready     4h\\n'  # E501 +            # normalize formatting with previous return objects +            if results['results'].startswith('NAME'): +                nodes = [] +                # removing header line and trailing new line character of node lines +                for node_results in results['results'].split('\n')[1:-1]: +                    parts = node_results.split() +                    nodes.append({'name': parts[0], 'schedulable': parts[1] == 'Ready'}) +                results['nodes'] = nodes + +            return results + +        results = {} +        results['returncode'] = 0 +        results['changed'] = False +        results['nodes'] = nodes + +        return results + +    @staticmethod +    def run_ansible(params, check_mode): +        '''run the idempotent ansible code''' +        nconfig = ManageNodeConfig(params['kubeconfig'], +                                   {'node': {'value': params['node'], 'include': True}, +                                    'selector': {'value': params['selector'], 'include': True}, +                                    'pod_selector': {'value': params['pod_selector'], 'include': True}, +                                    'schedulable': {'value': params['schedulable'], 'include': True}, +                                    'list_pods': {'value': params['list_pods'], 'include': True}, +                                    'evacuate': {'value': params['evacuate'], 'include': True}, +                                    'dry_run': {'value': params['dry_run'], 'include': True}, +                                    'force': {'value': params['force'], 'include': True}, +                                    'grace_period': {'value': params['grace_period'], 'include': True}, +                                   }) + +        oadm_mn = ManageNode(nconfig) +        # Run the oadm manage-node commands +        results = None +        changed = False +        if params['schedulable'] != None: +            if check_mode: +                # schedulable returns results after the fact. +                # We need to redo how this works to support check_mode completely. +                return {'changed': True, 'msg': 'CHECK_MODE: would have called schedulable.'} +            results = oadm_mn.schedulable() +            if 'changed' not in results: +                changed = True + +        if params['evacuate']: +            results = oadm_mn.evacuate() +            changed = True +        elif params['list_pods']: +            results = oadm_mn.list_pods() + +        if not results or results['returncode'] != 0: +            return {'failed': True, 'msg': results} + +        return {'changed': changed, 'results': results, 'state': "present"} + +# -*- -*- -*- End included fragment: class/oadm_manage_node.py -*- -*- -*- + +# -*- -*- -*- Begin included fragment: ansible/oadm_manage_node.py -*- -*- -*- + + +def main(): +    ''' +    ansible oadm module for manage-node +    ''' + +    module = AnsibleModule( +        argument_spec=dict( +            debug=dict(default=False, type='bool'), +            kubeconfig=dict(default='/etc/origin/master/admin.kubeconfig', type='str'), +            node=dict(default=None, type='list'), +            selector=dict(default=None, type='str'), +            pod_selector=dict(default=None, type='str'), +            schedulable=dict(default=None, type='bool'), +            list_pods=dict(default=False, type='bool'), +            evacuate=dict(default=False, type='bool'), +            dry_run=dict(default=False, type='bool'), +            force=dict(default=False, type='bool'), +            grace_period=dict(default=None, type='int'), +        ), +        mutually_exclusive=[["selector", "node"], ['evacuate', 'list_pods'], ['list_pods', 'schedulable']], +        required_one_of=[["node", "selector"]], + +        supports_check_mode=True, +    ) +    results = ManageNode.run_ansible(module.params, module.check_mode) + +    if 'failed' in results: +        module.fail_json(**results) + +    module.exit_json(**results) + + +if __name__ == "__main__": +    main() + +# -*- -*- -*- End included fragment: ansible/oadm_manage_node.py -*- -*- -*- diff --git a/roles/lib_openshift/library/oc_edit.py b/roles/lib_openshift/library/oc_edit.py index ad158e56a..064229d63 100644 --- a/roles/lib_openshift/library/oc_edit.py +++ b/roles/lib_openshift/library/oc_edit.py @@ -947,7 +947,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/library/oc_obj.py b/roles/lib_openshift/library/oc_obj.py index b5ccfe700..132752f47 100644 --- a/roles/lib_openshift/library/oc_obj.py +++ b/roles/lib_openshift/library/oc_obj.py @@ -926,7 +926,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/library/oc_route.py b/roles/lib_openshift/library/oc_route.py index fb51367fc..15b184083 100644 --- a/roles/lib_openshift/library/oc_route.py +++ b/roles/lib_openshift/library/oc_route.py @@ -951,7 +951,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/library/oc_scale.py b/roles/lib_openshift/library/oc_scale.py index 1259c5711..705b05dc8 100644 --- a/roles/lib_openshift/library/oc_scale.py +++ b/roles/lib_openshift/library/oc_scale.py @@ -901,7 +901,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/library/oc_secret.py b/roles/lib_openshift/library/oc_secret.py index 69dcb314b..f5e994b27 100644 --- a/roles/lib_openshift/library/oc_secret.py +++ b/roles/lib_openshift/library/oc_secret.py @@ -947,7 +947,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/library/oc_version.py b/roles/lib_openshift/library/oc_version.py index d74564352..dbd398c84 100644 --- a/roles/lib_openshift/library/oc_version.py +++ b/roles/lib_openshift/library/oc_version.py @@ -871,7 +871,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/src/ansible/oadm_manage_node.py b/roles/lib_openshift/src/ansible/oadm_manage_node.py new file mode 100644 index 000000000..b870c1211 --- /dev/null +++ b/roles/lib_openshift/src/ansible/oadm_manage_node.py @@ -0,0 +1,38 @@ +# pylint: skip-file +# flake8: noqa + + +def main(): +    ''' +    ansible oadm module for manage-node +    ''' + +    module = AnsibleModule( +        argument_spec=dict( +            debug=dict(default=False, type='bool'), +            kubeconfig=dict(default='/etc/origin/master/admin.kubeconfig', type='str'), +            node=dict(default=None, type='list'), +            selector=dict(default=None, type='str'), +            pod_selector=dict(default=None, type='str'), +            schedulable=dict(default=None, type='bool'), +            list_pods=dict(default=False, type='bool'), +            evacuate=dict(default=False, type='bool'), +            dry_run=dict(default=False, type='bool'), +            force=dict(default=False, type='bool'), +            grace_period=dict(default=None, type='int'), +        ), +        mutually_exclusive=[["selector", "node"], ['evacuate', 'list_pods'], ['list_pods', 'schedulable']], +        required_one_of=[["node", "selector"]], + +        supports_check_mode=True, +    ) +    results = ManageNode.run_ansible(module.params, module.check_mode) + +    if 'failed' in results: +        module.fail_json(**results) + +    module.exit_json(**results) + + +if __name__ == "__main__": +    main() diff --git a/roles/lib_openshift/src/class/oadm_manage_node.py b/roles/lib_openshift/src/class/oadm_manage_node.py new file mode 100644 index 000000000..61b6a5ebe --- /dev/null +++ b/roles/lib_openshift/src/class/oadm_manage_node.py @@ -0,0 +1,209 @@ +# pylint: skip-file +# flake8: noqa + + +class ManageNodeException(Exception): +    ''' manage-node exception class ''' +    pass + + +class ManageNodeConfig(OpenShiftCLIConfig): +    ''' ManageNodeConfig is a DTO for the manage-node command.''' +    def __init__(self, kubeconfig, node_options): +        super(ManageNodeConfig, self).__init__(None, None, kubeconfig, node_options) + + +# pylint: disable=too-many-instance-attributes +class ManageNode(OpenShiftCLI): +    ''' Class to wrap the oc command line tools ''' + +    # pylint allows 5 +    # pylint: disable=too-many-arguments +    def __init__(self, +                 config, +                 verbose=False): +        ''' Constructor for ManageNode ''' +        super(ManageNode, self).__init__(None, config.kubeconfig) +        self.config = config + +    def evacuate(self): +        ''' formulate the params and run oadm manage-node ''' +        return self._evacuate(node=self.config.config_options['node']['value'], +                              selector=self.config.config_options['selector']['value'], +                              pod_selector=self.config.config_options['pod_selector']['value'], +                              dry_run=self.config.config_options['dry_run']['value'], +                              grace_period=self.config.config_options['grace_period']['value'], +                              force=self.config.config_options['force']['value'], +                             ) +    def get_nodes(self, node=None, selector=''): +        '''perform oc get node''' +        _node = None +        _sel = None +        if node: +            _node = node +        if selector: +            _sel = selector + +        results = self._get('node', rname=_node, selector=_sel) +        if results['returncode'] != 0: +            return results + +        nodes = [] +        items = None +        if results['results'][0]['kind'] == 'List': +            items = results['results'][0]['items'] +        else: +            items = results['results'] + +        for node in items: +            _node = {} +            _node['name'] = node['metadata']['name'] +            _node['schedulable'] = True +            if 'unschedulable' in node['spec']: +                _node['schedulable'] = False +            nodes.append(_node) + +        return nodes + +    def get_pods_from_node(self, node, pod_selector=None): +        '''return pods for a node''' +        results = self._list_pods(node=[node], pod_selector=pod_selector) + +        if results['returncode'] != 0: +            return results + +        # When a selector or node is matched it is returned along with the json. +        # We are going to split the results based on the regexp and then +        # load the json for each matching node. +        # Before we return we are going to loop over the results and pull out the node names. +        # {'node': [pod, pod], 'node': [pod, pod]} +        # 3.2 includes the following lines in stdout: "Listing matched pods on node:" +        all_pods = [] +        if "Listing matched" in results['results']: +            listing_match = re.compile('\n^Listing matched.*$\n', flags=re.MULTILINE) +            pods = listing_match.split(results['results']) +            for pod in pods: +                if pod: +                    all_pods.extend(json.loads(pod)['items']) + +        # 3.3 specific +        else: +            # this is gross but I filed a bug... +            # https://bugzilla.redhat.com/show_bug.cgi?id=1381621 +            # build our own json from the output. +            all_pods = json.loads(results['results'])['items'] + +        return all_pods + +    def list_pods(self): +        ''' run oadm manage-node --list-pods''' +        _nodes = self.config.config_options['node']['value'] +        _selector = self.config.config_options['selector']['value'] +        _pod_selector = self.config.config_options['pod_selector']['value'] + +        if not _nodes: +            _nodes = self.get_nodes(selector=_selector) +        else: +            _nodes = [{'name': name} for name in _nodes] + +        all_pods = {} +        for node in _nodes: +            results = self.get_pods_from_node(node['name'], pod_selector=_pod_selector) +            if isinstance(results, dict): +                return results +            all_pods[node['name']] = results + +        results = {} +        results['nodes'] = all_pods +        results['returncode'] = 0 +        return results + +    def schedulable(self): +        '''oadm manage-node call for making nodes unschedulable''' +        nodes = self.config.config_options['node']['value'] +        selector = self.config.config_options['selector']['value'] + +        if not nodes: +            nodes = self.get_nodes(selector=selector) +        else: +            tmp_nodes = [] +            for name in nodes: +                tmp_result = self.get_nodes(name) +                if isinstance(tmp_result, dict): +                    tmp_nodes.append(tmp_result) +                    continue +                tmp_nodes.extend(tmp_result) +            nodes = tmp_nodes + +        # This is a short circuit based on the way we fetch nodes. +        # If node is a dict/list then we've already fetched them. +        for node in nodes: +            if isinstance(node, dict) and 'returncode' in node: +                return {'results': nodes, 'returncode': node['returncode']} +            if isinstance(node, list) and 'returncode' in node[0]: +                return {'results': nodes, 'returncode': node[0]['returncode']} +        # check all the nodes that were returned and verify they are: +        # node['schedulable'] == self.config.config_options['schedulable']['value'] +        if any([node['schedulable'] != self.config.config_options['schedulable']['value'] for node in nodes]): + +            results = self._schedulable(node=self.config.config_options['node']['value'], +                                        selector=self.config.config_options['selector']['value'], +                                        schedulable=self.config.config_options['schedulable']['value']) + +            # 'NAME                            STATUS    AGE\\nip-172-31-49-140.ec2.internal   Ready     4h\\n'  # E501 +            # normalize formatting with previous return objects +            if results['results'].startswith('NAME'): +                nodes = [] +                # removing header line and trailing new line character of node lines +                for node_results in results['results'].split('\n')[1:-1]: +                    parts = node_results.split() +                    nodes.append({'name': parts[0], 'schedulable': parts[1] == 'Ready'}) +                results['nodes'] = nodes + +            return results + +        results = {} +        results['returncode'] = 0 +        results['changed'] = False +        results['nodes'] = nodes + +        return results + +    @staticmethod +    def run_ansible(params, check_mode): +        '''run the idempotent ansible code''' +        nconfig = ManageNodeConfig(params['kubeconfig'], +                                   {'node': {'value': params['node'], 'include': True}, +                                    'selector': {'value': params['selector'], 'include': True}, +                                    'pod_selector': {'value': params['pod_selector'], 'include': True}, +                                    'schedulable': {'value': params['schedulable'], 'include': True}, +                                    'list_pods': {'value': params['list_pods'], 'include': True}, +                                    'evacuate': {'value': params['evacuate'], 'include': True}, +                                    'dry_run': {'value': params['dry_run'], 'include': True}, +                                    'force': {'value': params['force'], 'include': True}, +                                    'grace_period': {'value': params['grace_period'], 'include': True}, +                                   }) + +        oadm_mn = ManageNode(nconfig) +        # Run the oadm manage-node commands +        results = None +        changed = False +        if params['schedulable'] != None: +            if check_mode: +                # schedulable returns results after the fact. +                # We need to redo how this works to support check_mode completely. +                return {'changed': True, 'msg': 'CHECK_MODE: would have called schedulable.'} +            results = oadm_mn.schedulable() +            if 'changed' not in results: +                changed = True + +        if params['evacuate']: +            results = oadm_mn.evacuate() +            changed = True +        elif params['list_pods']: +            results = oadm_mn.list_pods() + +        if not results or results['returncode'] != 0: +            return {'failed': True, 'msg': results} + +        return {'changed': changed, 'results': results, 'state': "present"} diff --git a/roles/lib_openshift/src/doc/manage_node b/roles/lib_openshift/src/doc/manage_node new file mode 100644 index 000000000..382377f3e --- /dev/null +++ b/roles/lib_openshift/src/doc/manage_node @@ -0,0 +1,88 @@ +# flake8: noqa +# pylint: skip-file + +DOCUMENTATION = ''' +--- +module: oadm_manage_node +short_description: Module to manage openshift nodes +description: +  - Manage openshift nodes programmatically. +options: +  kubeconfig: +    description: +    - The path for the kubeconfig file to use for authentication +    required: false +    default: /etc/origin/master/admin.kubeconfig +    aliases: [] +  debug: +    description: +    - Turn on debug output. +    required: false +    default: False +    aliases: [] +  node: +    description: +    - A list of the nodes being managed +    required: false +    default: None +    aliases: [] +  selector: +    description: +    - The selector when filtering on node labels +    required: false +    default: None +    aliases: [] +  pod_selector: +    description: +    - A selector when filtering on pod labels. +    required: false +    default: None +    aliases: [] +  evacuate: +    description: +    - Remove all pods from a node. +    required: false +    default: False +    aliases: [] +  schedulable: +    description: +    - whether or not openshift can schedule pods on this node +    required: False +    default: None +    aliases: [] +  dry_run: +    description: +    - This shows the pods that would be migrated if evacuate were called +    required: False +    default: False +    aliases: [] +  grace_period: +    description: +    - Grace period (seconds) for pods being deleted. +    required: false +    default: None +    aliases: [] +  force: +    description: +    - Whether or not to attempt to force this action in openshift +    required: false +    default: None +    aliases: [] +author: +- "Kenny Woodson <kwoodson@redhat.com>" +extends_documentation_fragment: [] +''' + +EXAMPLES = ''' +- name: oadm manage-node --schedulable=true --selector=ops_node=new +  oadm_manage_node: +    selector: ops_node=new +    schedulable: True +  register: schedout + +- name: oadm manage-node my-k8s-node-5 --evacuate +  oadm_manage_node: +    node:  my-k8s-node-5 +    evacuate: True +    force: True +''' diff --git a/roles/lib_openshift/src/lib/base.py b/roles/lib_openshift/src/lib/base.py index 8b5491d6b..2656f572e 100644 --- a/roles/lib_openshift/src/lib/base.py +++ b/roles/lib_openshift/src/lib/base.py @@ -216,7 +216,7 @@ class OpenShiftCLI(object):          if self.all_namespaces:              cmds.extend(['--all-namespaces']) -        elif self.namespace: +        elif self.namespace is not None and self.namespace.lower() not in ['none', 'emtpy']:  # E501              cmds.extend(['-n', self.namespace])          cmds.extend(cmd) diff --git a/roles/lib_openshift/src/sources.yml b/roles/lib_openshift/src/sources.yml index 5afcdc55d..b0835784b 100644 --- a/roles/lib_openshift/src/sources.yml +++ b/roles/lib_openshift/src/sources.yml @@ -1,4 +1,13 @@  --- +oadm_manage_node.py: +- doc/generated +- doc/license +- lib/import.py +- doc/manage_node +- ../../lib_utils/src/class/yedit.py +- lib/base.py +- class/oadm_manage_node.py +- ansible/oadm_manage_node.py  oc_edit.py:  - doc/generated  - doc/license diff --git a/roles/lib_openshift/src/test/integration/oadm_manage_node.yml b/roles/lib_openshift/src/test/integration/oadm_manage_node.yml new file mode 100755 index 000000000..69a701b17 --- /dev/null +++ b/roles/lib_openshift/src/test/integration/oadm_manage_node.yml @@ -0,0 +1,58 @@ +#!/usr/bin/ansible-playbook --module-path=../../../library/ +# ./oadm_manage_node.yml -M ../../../library -e "cli_master_test=$OPENSHIFT_MASTER cli_node_test=$OPENSHIFT_NODE +--- +- hosts: "{{ cli_master_test }}" +  gather_facts: no +  user: root +  tasks: +  - name: list pods from a node +    oadm_manage_node: +      list_pods: True +      node: +      - "{{ cli_node_test }}" +    register: podout +  - debug: var=podout + +  - assert: +      that: "'{{ cli_node_test }}' in podout.results.nodes" +      msg: Pod data was not returned + +  - name: set node to unschedulable +    oadm_manage_node: +      schedulable: False +      node: +      - "{{ cli_node_test }}" +    register: nodeout +  - debug: var=nodeout + +  - name: assert that schedulable=False +    assert: +      that: nodeout.results.nodes[0]['schedulable'] == False +      msg: "{{ cli_node_test }} schedulable set to True" + +  - name: get node scheduable +    oc_obj: +      kind: node +      state: list +      name: "{{ cli_node_test }}" +      namespace: None +    register: nodeout + +  - debug: var=nodeout + +  - name: assert that schedulable=False +    assert: +      that: nodeout.results.results[0]['spec']['unschedulable'] + +  - name: set node to schedulable +    oadm_manage_node: +      schedulable: True +      node: +      - "{{ cli_node_test }}" +    register: nodeout +  - debug: var=nodeout + +  - name: assert that schedulable=False +    assert: +      that: nodeout.results.nodes[0]['schedulable'] +      msg: "{{ cli_node_test }} schedulable set to False" diff --git a/roles/lib_openshift/src/test/unit/oadm_manage_node.py b/roles/lib_openshift/src/test/unit/oadm_manage_node.py new file mode 100755 index 000000000..8fd6f9c55 --- /dev/null +++ b/roles/lib_openshift/src/test/unit/oadm_manage_node.py @@ -0,0 +1,177 @@ +#!/usr/bin/env python2 +''' + Unit tests for oadm_manage_node +''' +# To run +# python -m unittest version +# +# . +# Ran 2 tests in 0.001s +# +# OK + +import os +import sys +import unittest +import mock + +# Removing invalid variable names for tests so that I can +# keep them brief +# pylint: disable=invalid-name,no-name-in-module +# Disable import-error b/c our libraries aren't loaded in jenkins +# pylint: disable=import-error +# place class in our python path +module_path = os.path.join('/'.join(os.path.realpath(__file__).split('/')[:-4]), 'library')  # noqa: E501 +sys.path.insert(0, module_path) +from oadm_manage_node import ManageNode  # noqa: E402 + + +class ManageNodeTest(unittest.TestCase): +    ''' +     Test class for oadm_manage_node +    ''' + +    def setUp(self): +        ''' setup method will create a file and set to known configuration ''' +        pass + +    @mock.patch('oadm_manage_node.ManageNode.openshift_cmd') +    def test_list_pods(self, mock_openshift_cmd): +        ''' Testing a get ''' +        params = {'node': ['ip-172-31-49-140.ec2.internal'], +                  'schedulable': None, +                  'selector': None, +                  'pod_selector': None, +                  'list_pods': True, +                  'kubeconfig': '/etc/origin/master/admin.kubeconfig', +                  'evacuate': False, +                  'grace_period': False, +                  'dry_run': False, +                  'force': False} + +        pod_list = '''{ +    "metadata": {}, +    "items": [ +        { +            "metadata": { +                "name": "docker-registry-1-xuhik", +                "generateName": "docker-registry-1-", +                "namespace": "default", +                "selfLink": "/api/v1/namespaces/default/pods/docker-registry-1-xuhik", +                "uid": "ae2a25a2-e316-11e6-80eb-0ecdc51fcfc4", +                "resourceVersion": "1501", +                "creationTimestamp": "2017-01-25T15:55:23Z", +                "labels": { +                    "deployment": "docker-registry-1", +                    "deploymentconfig": "docker-registry", +                    "docker-registry": "default" +                }, +                "annotations": { +                    "openshift.io/deployment-config.latest-version": "1", +                    "openshift.io/deployment-config.name": "docker-registry", +                    "openshift.io/deployment.name": "docker-registry-1", +                    "openshift.io/scc": "restricted" +                } +            }, +            "spec": {} +        }, +        { +            "metadata": { +                "name": "router-1-kp3m3", +                "generateName": "router-1-", +                "namespace": "default", +                "selfLink": "/api/v1/namespaces/default/pods/router-1-kp3m3", +                "uid": "9e71f4a5-e316-11e6-80eb-0ecdc51fcfc4", +                "resourceVersion": "1456", +                "creationTimestamp": "2017-01-25T15:54:56Z", +                "labels": { +                    "deployment": "router-1", +                    "deploymentconfig": "router", +                    "router": "router" +                }, +                "annotations": { +                    "openshift.io/deployment-config.latest-version": "1", +                    "openshift.io/deployment-config.name": "router", +                    "openshift.io/deployment.name": "router-1", +                    "openshift.io/scc": "hostnetwork" +                } +            }, +            "spec": {} +        }] +}''' + +        mock_openshift_cmd.side_effect = [ +            {"cmd": "/usr/bin/oadm manage-node ip-172-31-49-140.ec2.internal --list-pods", +             "results": pod_list, +             "returncode": 0} +        ] + +        results = ManageNode.run_ansible(params, False) + +        # returned a single node +        self.assertTrue(len(results['results']['nodes']) == 1) +        # returned 2 pods +        self.assertTrue(len(results['results']['nodes']['ip-172-31-49-140.ec2.internal']) == 2) + +    @mock.patch('oadm_manage_node.ManageNode.openshift_cmd') +    def test_schedulable_false(self, mock_openshift_cmd): +        ''' Testing a get ''' +        params = {'node': ['ip-172-31-49-140.ec2.internal'], +                  'schedulable': False, +                  'selector': None, +                  'pod_selector': None, +                  'list_pods': False, +                  'kubeconfig': '/etc/origin/master/admin.kubeconfig', +                  'evacuate': False, +                  'grace_period': False, +                  'dry_run': False, +                  'force': False} + +        node = [{ +            "apiVersion": "v1", +            "kind": "Node", +            "metadata": { +                "creationTimestamp": "2017-01-26T14:34:43Z", +                "labels": { +                    "beta.kubernetes.io/arch": "amd64", +                    "beta.kubernetes.io/instance-type": "m4.large", +                    "beta.kubernetes.io/os": "linux", +                    "failure-domain.beta.kubernetes.io/region": "us-east-1", +                    "failure-domain.beta.kubernetes.io/zone": "us-east-1c", +                    "hostname": "opstest-node-compute-0daaf", +                    "kubernetes.io/hostname": "ip-172-31-51-111.ec2.internal", +                    "ops_node": "old", +                    "region": "us-east-1", +                    "type": "compute" +                }, +                "name": "ip-172-31-51-111.ec2.internal", +                "resourceVersion": "6936", +                "selfLink": "/api/v1/nodes/ip-172-31-51-111.ec2.internal", +                "uid": "93d7fdfb-e3d4-11e6-a982-0e84250fc302" +            }, +            "spec": { +                "externalID": "i-06bb330e55c699b0f", +                "providerID": "aws:///us-east-1c/i-06bb330e55c699b0f", +            }}] + +        mock_openshift_cmd.side_effect = [ +            {"cmd": "/usr/bin/oc get node -o json ip-172-31-49-140.ec2.internal", +             "results": node, +             "returncode": 0}, +            {"cmd": "/usr/bin/oadm manage-node ip-172-31-49-140.ec2.internal --schedulable=False", +             "results": "NAME                            STATUS    AGE\n" + +                        "ip-172-31-49-140.ec2.internal   Ready,SchedulingDisabled     5h\n", +             "returncode": 0}] +        results = ManageNode.run_ansible(params, False) + +        self.assertTrue(results['changed']) +        self.assertEqual(results['results']['nodes'][0]['name'], 'ip-172-31-49-140.ec2.internal') +        self.assertEqual(results['results']['nodes'][0]['schedulable'], False) + +    def tearDown(self): +        '''TearDown method''' +        pass + + +if __name__ == "__main__": +    unittest.main() | 
