Source code for sos.report.plugins

# Copyright (C) 2006 Steve Conklin <sconklin@redhat.com>

# This file is part of the sos project: https://github.com/sosreport/sos
#
# This copyrighted material is made available to anyone wishing to use,
# modify, copy, or redistribute it subject to the terms and conditions of
# version 2 of the GNU General Public License.
#
# See the LICENSE file in the source distribution for further information.

""" This exports methods available for use by plugins for sos """

from sos.utilities import (sos_get_command_output, import_module, grep,
                           fileobj, tail, is_executable)
import os
import glob
import re
import stat
from time import time
import logging
import fnmatch
import errno

from datetime import datetime


[docs]def regex_findall(regex, fname): """Return a list of all non overlapping matches in the string(s)""" try: with fileobj(fname) as f: return re.findall(regex, f.read(), re.MULTILINE) except AttributeError: return []
def _mangle_command(command, name_max): mangledname = re.sub(r"^/(usr/|)(bin|sbin)/", "", command) mangledname = re.sub(r"[^\w\-\.\/]+", "_", mangledname) mangledname = re.sub(r"/", ".", mangledname).strip(" ._-") mangledname = mangledname[0:name_max] return mangledname def _path_in_path_list(path, path_list): return any((p == path or path.startswith(os.path.abspath(p)+os.sep) for p in path_list)) def _node_type(st): """ return a string indicating the type of special node represented by the stat buffer st (block, character, fifo, socket). """ _types = [ (stat.S_ISBLK, "block device"), (stat.S_ISCHR, "character device"), (stat.S_ISFIFO, "named pipe"), (stat.S_ISSOCK, "socket") ] for t in _types: if t[0](st.st_mode): return t[1] def _file_is_compressed(path): """Check if a file appears to be compressed Return True if the file specified by path appears to be compressed, or False otherwise by testing the file name extension against a list of known file compression extentions. """ return path.endswith(('.gz', '.xz', '.bz', '.bz2')) _certmatch = re.compile("-*BEGIN.*?-*END", re.DOTALL) _cert_replace = "-----SCRUBBED"
[docs]class SoSPredicate(object): """A class to implement collection predicates. A predicate gates the collection of data by an sos plugin. For any `add_cmd_output()`, `add_copy_spec()` or `add_journal()` call, the passed predicate will be evaulated and collection will proceed if the result is `True`, and not otherwise. Predicates may be used to control conditional data collection without the need for explicit conditional blocks in plugins. :param owner: The ``Plugin`` object creating the predicate :type owner: ``Plugin`` :param dry_run: Is sos running in dry_run mode? :type dry_run: ``bool`` :param kmods: Kernel module name(s) to check presence of :type kmods: ``list``, or ``str`` of single name :param services: Service name(s) to check if running :type services: ``list``, or ``str`` of single name :param packages: Package name(s) to check presence of :type packages: ``list``, or ``str`` of single name :param cmd_outputs: Command to run, with output string to check :type cmd_outputs: ``list`` of ``dict``s, or single ``dict`` taking form {'cmd': <command to run>, 'output': <string that output should contain>} :param arch: Architecture(s) that the local system is matched against :type arch: ``list``, or ``str`` of single architecture :param required: For each parameter provided, should the checks require all items, no items, or any items provided :type required: ``dict``, with keys matching parameter names and values being either 'any', 'all', or 'none. Default 'any'. """ #: The plugin that owns this predicate _owner = None #: Skip all collection? dry_run = False #: Kernel module enablement list kmods = [] #: Services enablement list services = [] #: Package presence list packages = [] # Command output inclusion pairs {'cmd': 'foo --help', 'output': 'bar'} cmd_outputs = [] #: Allowed architecture(s) of the system arch = [] def __str(self, quote=False, prefix="", suffix=""): """Return a string representation of this SoSPredicate with optional prefix, suffix and value quoting. """ quotes = '"%s"' pstr = "dry_run=%s, " % self.dry_run kmods = self.kmods kmods = [quotes % k for k in kmods] if quote else kmods pstr += "kmods=[%s], " % (",".join(kmods)) services = self.services services = [quotes % s for s in services] if quote else services pstr += "services=[%s], " % (",".join(services)) pkgs = self.packages pkgs = [quotes % p for p in pkgs] if quote else pkgs pstr += "packages=[%s], " % (",".join(pkgs)) cmdoutputs = [ "{ %s: %s, %s: %s }" % (quotes % "cmd", quotes % cmdoutput['cmd'], quotes % "output", quotes % cmdoutput['output']) for cmdoutput in self.cmd_outputs ] pstr += "cmdoutputs=[%s], " % (",".join(cmdoutputs)) arches = self.arch arches = [quotes % a for a in arches] if quote else arches pstr += "arches=[%s]" % (",".join(arches)) return prefix + pstr + suffix def __str__(self): """Return a string representation of this SoSPredicate. "dry_run=False, kmods=[], services=[], cmdoutputs=[]" """ return self.__str() def __repr__(self): """Return a machine readable string representation of this SoSPredicate. "SoSPredicate(dry_run=False, kmods=[], services=[], cmdoutputs=[])" """ return self.__str(quote=True, prefix="SoSPredicate(", suffix=")") def _check_required_state(self, items, required): """Helper to simplify checking the state of the predicate's evaluations against the setting of the required state of that evaluation """ if required == 'any': return any(items) elif required == 'all': return all(items) elif required == 'none': return not any(items) def _failed_or_forbidden(self, test, item): """Helper to direct failed predicates to provide the proper messaging based on the required check type :param test: The type of check we're doing, e.g. kmods, arch :param item: The string of what failed """ _req = self.required[test] if _req != 'none': self._failed[test].append(item) else: self._forbidden[test].append(item) def _eval_kmods(self): if not self.kmods or self._owner.get_option('allow_system_changes'): return True _kmods = [] # Are kernel modules loaded? for kmod in self.kmods: res = self._owner.is_module_loaded(kmod) _kmods.append(res) if not res: self._failed_or_forbidden('kmods', kmod) return self._check_required_state(_kmods, self.required['kmods']) def _eval_services(self): if not self.services: return True _svcs = [] for svc in self.services: res = self._owner.is_service_running(svc) _svcs.append(res) if not res: self._failed_or_forbidden('services', svc) return self._check_required_state(_svcs, self.required['services']) def _eval_packages(self): if not self.packages: return True _pkgs = [] for pkg in self.packages: res = self._owner.is_installed(pkg) _pkgs.append(res) if not res: self._failed_or_forbidden('packages', pkg) return self._check_required_state(_pkgs, self.required['packages']) def _eval_cmd_output(self, cmd_output): """Does 'cmd' output contain string 'output'?""" if 'cmd' not in cmd_output or 'output' not in cmd_output: return False result = sos_get_command_output(cmd_output['cmd']) if result['status'] != 0: return False for line in result['output'].splitlines(): if cmd_output['output'] in line: return True return False def _eval_cmd_outputs(self): if not self.cmd_outputs: return True _cmds = [] for cmd in self.cmd_outputs: res = self._eval_cmd_output(cmd) _cmds.append(res) if not res: self._failed_or_forbidden( 'cmd_outputs', "%s: %s" % (cmd['cmd'], cmd['output']) ) return self._check_required_state(_cmds, self.required['cmd_outputs']) def _eval_arch(self): if not self.arch: return True # a test for 'all' against arch does not make sense, so only test to # see if the system's reported architecture is in the last of 'allowed' # arches requested by the predicate _arch = self._owner.policy.get_arch() regex = '(?:%s)' % '|'.join(self.arch) if self.required['arch'] == 'none': if re.match(regex, _arch): self._forbidden['architecture'].append(_arch) return False return True if re.match(regex, _arch): return True self._failed['architecture'].append(_arch) return False def _report_failed(self): """Return a string informing user what caused the predicate to fail evaluation """ msg = '' _substr = "required %s missing: %s." for key, val in self._failed.items(): if not val: continue val = set(val) msg += _substr % (key, ', '.join(v for v in val)) return msg def _report_forbidden(self): """Return a string informing the user that a forbidden condition exists which caused the predicate to fail """ msg = '' _substr = "forbidden %s '%s' found." for key, val in self._forbidden.items(): if not val: continue val = set(val) msg += _substr % (key, ', '.join(v for v in val)) return msg
[docs] def report_failure(self): """Used by `Plugin()` to obtain the error string based on if the reason was a failed check or a forbidden check """ msg = [self._report_failed(), self._report_forbidden()] return " ".join(msg).lstrip()
def __nonzero__(self): """Predicate evaluation hook. """ # Null predicate? if not any([self.kmods, self.services, self.packages, self.cmd_outputs, self.arch, self.dry_run]): return True return ((self._eval_kmods() and self._eval_services() and self._eval_packages() and self._eval_cmd_outputs() and self._eval_arch()) and not self.dry_run) def __bool__(self): # Py3 evaluation ends in a __bool__() call where py2 ends in a call # to __nonzero__(). Wrap the latter here, to support both versions return self.__nonzero__() def __init__(self, owner, dry_run=False, kmods=[], services=[], packages=[], cmd_outputs=[], arch=[], required={}): """Initialise a new SoSPredicate object """ self._owner = owner self.kmods = list(kmods) self.services = list(services) self.packages = list(packages) self.arch = list(arch) if not isinstance(cmd_outputs, list): cmd_outputs = [cmd_outputs] self.cmd_outputs = cmd_outputs self.dry_run = dry_run | self._owner.commons['cmdlineopts'].dry_run self.required = {'kmods': 'any', 'services': 'any', 'packages': 'any', 'cmd_outputs': 'any', 'arch': 'any'} self.required.update({ k: v for k, v in required.items() if required[k] != self.required[k] }) #: Dict holding failed evaluations self._failed = { 'kmods': [], 'services': [], 'packages': [], 'cmd_outputs': [], 'architecture': [] } self._forbidden = { 'kmods': [], 'services': [], 'packages': [], 'cmd_outputs': [], 'architecture': [] }
[docs]class SoSCommand(object): """A class to represent a command to be collected. A SoSCommand() object is instantiated for each command handed to an _add_cmd_output() call, so that we no longer need to pass around a very long tuple to handle the parameters. Any option supported by _add_cmd_output() is passed to the SoSCommand object and converted to an attribute. SoSCommand.__dict__ is then passed to _get_command_output_now() for each command to be collected. """ def __init__(self, **kwargs): self.__dict__.update(kwargs) def __str__(self): """Return a human readable string representation of this SoSCommand """ return ', '.join("%s=%r" % (param, val) for (param, val) in sorted(self.__dict__.items()))
[docs]class Plugin(object): """This is the base class for sosreport plugins. Plugins should subclass this and set the class variables where applicable. :param commons: A set of information that is shared internally so that plugins may access the same dataset. This is provided automatically by sos :type commons: ``dict`` Each `Plugin()` subclass should also subclass at least one tagging class, e.g. ``RedHatPlugin``, to support that distribution. If different distributions require different collections, each distribution should have its own subclass of the Plugin that also subclasses the tagging class for their respective distributions. :cvar plugin_name: The name of the plugin, will be returned by `name()` :vartype plugin_name: ``str`` :cvar version: The version of the plugin, defaults to 'unversioned' :vartype version: ``str`` :cvar packages: Package name(s) that, if installed, enable this plugin :vartype packages: ``tuple`` :cvar files: File path(s) that, if present, enable this plugin :vartype files: ``tuple`` :cvar commands: Executables that, if present, enable this plugin :vartype commands: ``tuple`` :cvar kernel_mods: Kernel module(s) that, if loaded, enable this plugin :vartype kernel_mods: ``tuple`` :cvar services: Service name(s) that, if running, enable this plugin :vartype services: ``tuple`` :cvar architectures: Architecture(s) this plugin is enabled for. Defaults to 'none' to enable on all arches. :vartype architectures: ``tuple``, or ``None`` :cvar profiles: Name(s) of profile(s) this plugin belongs to :vartype profiles: ``tuple`` :cvar plugin_timeout: Timeout in seconds for this plugin as a whole :vartype plugin_timeout: ``int`` :cvar cmd_timeout: Timeout in seconds for individual commands :vartype cmd_timeout: ``int`` """ plugin_name = None version = 'unversioned' packages = () files = () commands = () kernel_mods = () services = () architectures = None archive = None profiles = () sysroot = '/' plugin_timeout = 300 cmd_timeout = 300 _timeout_hit = False cmdtags = {} filetags = {} # Default predicates predicate = None cmd_predicate = None _default_plug_opts = [ ('timeout', 'Timeout in seconds for plugin', 'fast', -1), ('postproc', 'Enable post-processing collected plugin data', 'fast', True) ] def __init__(self, commons): if not getattr(self, "option_list", False): self.option_list = [] self.copied_files = [] self.executed_commands = [] self._env_vars = set() self.alerts = [] self.custom_text = "" self.opt_names = [] self.opt_parms = [] self.commons = commons self.forbidden_paths = [] self.copy_paths = set() self.copy_strings = [] self.collect_cmds = [] self.sysroot = commons['sysroot'] self.policy = commons['policy'] self.devices = commons['devices'] self.manifest = None self.soslog = self.commons['soslog'] if 'soslog' in self.commons \ else logging.getLogger('sos') # add the default plugin opts self.option_list.extend(self._default_plug_opts) # get the option list into a dictionary for opt in self.option_list: self.opt_names.append(opt[0]) self.opt_parms.append({'desc': opt[1], 'speed': opt[2], 'enabled': opt[3]}) # Initialise the default --dry-run predicate self.set_predicate(SoSPredicate(self))
[docs] def set_plugin_manifest(self, manifest): """Pass in a manifest object to the plugin to write to :param manifest: The manifest that the plugin will add metadata to :type manifest: ``SoSManifest`` """ self.manifest = manifest # add these here for organization when they actually get set later self.manifest.add_field('start_time', '') self.manifest.add_field('end_time', '') self.manifest.add_field('run_time', '') self.manifest.add_field('setup_start', '') self.manifest.add_field('setup_end', '') self.manifest.add_field('setup_time', '') self.manifest.add_field('timeout_hit', False) self.manifest.add_list('commands', []) self.manifest.add_list('files', [])
@property def timeout(self): """Returns either the default plugin timeout value, the value as provided on the commandline via -k plugin.timeout=value, or the value of the global --plugin-timeout option. """ _timeout = None try: opt_timeout = self.get_option('plugin_timeout') own_timeout = int(self.get_option('timeout')) if opt_timeout is None: _timeout = own_timeout elif opt_timeout is not None and own_timeout == -1: _timeout = int(opt_timeout) elif opt_timeout is not None and own_timeout > -1: _timeout = own_timeout else: return None except ValueError: return self.plugin_timeout # Default to known safe value if _timeout is not None and _timeout > -1: return _timeout return self.plugin_timeout def set_timeout_hit(self): self._timeout_hit = True self.manifest.add_field('end_time', datetime.now()) self.manifest.add_field('timeout_hit', True)
[docs] def check_timeout(self): """ Checks to see if the plugin has hit its timeout. This is set when the sos.collect_plugin() method hits a timeout and terminates the thread. From there, a Popen() call can still continue to run, and we need to manually terminate it. Thus, check_timeout() should only be called in sos_get_command_output(). Since sos_get_command_output() is not plugin aware, this method is handed to that call to use as a polling method, to avoid passing the entire plugin object. :returns: ``True`` if timeout has been hit, else ``False`` :rtype: ``bool`` """ return self._timeout_hit
[docs] @classmethod def name(cls): """Get the name of the plugin :returns: The name of the plugin, in lowercase :rtype: ``str`` """ if cls.plugin_name: return cls.plugin_name return cls.__name__.lower()
def _format_msg(self, msg): return "[plugin:%s] %s" % (self.name(), msg) def _log_error(self, msg): self.soslog.error(self._format_msg(msg)) def _log_warn(self, msg): self.soslog.warning(self._format_msg(msg)) def _log_info(self, msg): self.soslog.info(self._format_msg(msg)) def _log_debug(self, msg): self.soslog.debug(self._format_msg(msg))
[docs] def join_sysroot(self, path): """Join a given path with the configured sysroot :param path: The filesystem path that needs to be joined :type path: ``str`` :returns: The joined filesystem path :rtype: ``str`` """ if path[0] == os.sep: path = path[1:] return os.path.join(self.sysroot, path)
[docs] def strip_sysroot(self, path): """Remove the configured sysroot from a filesystem path :param path: The filesystem path to strip sysroot from :type path: ``str`` :returns: The stripped filesystem path :rtype: ``str`` """ if not self.use_sysroot(): return path if path.startswith(self.sysroot): return path[len(self.sysroot):] return path
[docs] def use_sysroot(self): """Determine if the configured sysroot needs to be used :returns: ``True`` if sysroot is not `/`, else ``False`` :rtype: ``bool`` """ return self.sysroot != os.path.abspath(os.sep)
[docs] def tmp_in_sysroot(self): """Check if sysroot is within the archive's temp directory :returns: ``True`` if sysroot is in the archive's temp directory, else ``False`` :rtype: ``bool`` """ paths = [self.sysroot, self.archive.get_tmp_dir()] return os.path.commonprefix(paths) == self.sysroot
[docs] def is_installed(self, package_name): """Is the package $package_name installed? :param package_name: The name of the package to check :type package_name: ``str`` :returns: ``True`` id the package is installed, else ``False`` :rtype: ``bool`` """ return self.policy.pkg_by_name(package_name) is not None
[docs] def is_service(self, name): """Does the service $name exist on the system? :param name: The name of the service to check :type name: ``str`` :returns: ``True`` if service is present on the system, else ``False`` :rtype: ``bool`` """ return self.policy.init_system.is_service(name)
[docs] def is_service_enabled(self, name): """Is the service $name enabled? :param name: The name of the service to check :type name: ``str`` :returns: ``True if service is enabled on the system, else ``False`` :rtype: ``bool`` """ return self.policy.init_system.is_enabled(name)
[docs] def is_service_disabled(self, name): """Is the service $name disabled? :param name: The name of the service to check :type name: ``str`` :returns: ``True`` if service is disabled on the system, else ``False`` :rtype: ``bool`` """ return self.policy.init_system.is_disabled(name)
[docs] def is_service_running(self, name): """Is the service $name currently running? :param name: The name of the service to check :type name: ``str`` :returns: ``True`` if the service is running on the system, else ``False`` :rtype: ``bool`` """ return self.policy.init_system.is_running(name)
[docs] def get_service_status(self, name): """Return the reported status for service $name :param name: The name of the service to check :type name: ``str`` :returns: The state of the service according to the init system :rtype: ``str`` """ return self.policy.init_system.get_service_status(name)['status']
[docs] def get_service_names(self, regex): """Get all service names matching regex :param regex: A regex to match service names against :type regex: ``str`` :returns: All service name(s) matching the given `regex` :rtype: ``list`` """ return self.policy.init_system.get_service_names(regex)
[docs] def set_predicate(self, pred): """Set or clear the default predicate for this plugin. :param pred: The predicate to use as the default for this plugin :type pred: ``SoSPredicate`` """ self.predicate = pred
[docs] def set_cmd_predicate(self, pred): """Set or clear the default predicate for command collection for this plugin. If set, this predecate takes precedence over the `Plugin` default predicate for command and journal data collection. :param pred: The predicate to use as the default command predicate :type pred: ``SoSPredicate`` """ self.cmd_predicate = pred
[docs] def get_predicate(self, cmd=False, pred=None): """Get the current default `Plugin` or command predicate. :param cmd: If a command predicate is set, should it be used. :type cmd: ``bool`` :param pred: An optional predicate to pass if no command or plugin predicate is set :type pred: ``SoSPredicate`` :returns: `pred` if neither a command predicate or plugin predicate is set. The command predicate if one is set and `cmd` is ``True``, else the plugin default predicate (which may be ``None``). :rtype: ``SoSPredicate`` or ``None`` """ if pred is not None: return pred if cmd and self.cmd_predicate is not None: return self.cmd_predicate return self.predicate
[docs] def test_predicate(self, cmd=False, pred=None): """Test the current predicate and return its value. :param cmd: ``True`` if the predicate is gating a command or ``False`` otherwise. :param pred: An optional predicate to override the current ``Plugin`` or command predicate. :returns: ``True`` or ``False`` based on predicate evaluation, or ``False`` if no predicate :rtype: ``bool`` """ pred = self.get_predicate(cmd=cmd, pred=pred) if pred is not None: return bool(pred) return False
[docs] def log_skipped_cmd(self, pred, cmd, kmods=False, services=False, changes=False): """Log that a command was skipped due to predicate evaluation. Emit a warning message indicating that a command was skipped due to predicate evaluation. If ``kmods`` or ``services`` are ``True`` then the list of expected kernel modules or services will be included in the log message. If ``allow_changes`` is ``True`` a message indicating that the missing data can be collected by using the "--allow-system-changes" command line option will be included. :param pred: The predicate that caused the command to be skipped :type pred: ``SoSPredicate`` :param cmd: The command that was skipped :type cmd: ``str`` :param kmods: Did kernel modules cause the command to be skipped :type kmods: ``bool`` :param services: Did services cause the command to be skipped :type services: ``bool`` :param changes: Is the `--allow-system-changes` enabled :type changes: ``bool`` """ msg = "skipped command '%s': %s" % (cmd, pred.report_failure()) if changes: msg += " Use '--allow-system-changes' to enable collection." self._log_warn(msg)
[docs] def do_cmd_private_sub(self, cmd, desc=""): """Remove certificate and key output archived by sos report. Any matching instances are replaced with: '-----SCRUBBED' and this function does not take a regexp or substituting string. :param cmd: The name of the binary to scrub certificate output from :type cmd: ``str`` :param desc: An identifier to add to the `SCRUBBED` header line :type desc: ``str`` :returns: Number of replacements made :rtype: ``int`` """ if not self.executed_commands: return 0 self._log_debug("Scrubbing certs and keys for commands matching %s" % (cmd)) replace = "%s %s" % (_cert_replace, desc) if desc else _cert_replace return self.do_cmd_output_sub(cmd, _certmatch, replace)
[docs] def do_cmd_output_sub(self, cmd, regexp, subst): """Apply a regexp substitution to command output archived by sosreport. This is used to obfuscate sensitive information captured by command output collection via plugins. :param cmd: The command name/binary name for collected output that needs to be obfuscated. Internally globbed with a leading and trailing `*` :type cmd: ``str`` :param regexp: A regex to match the contents of the command output against :type regexp: ``str`` or compile ``re`` object :param subst: The substitution string used to replace matches from `regexp` :type subst: ``str`` :returns: Number of replacements made :rtype: ``int`` """ globstr = '*' + cmd + '*' pattern = regexp.pattern if hasattr(regexp, "pattern") else regexp self._log_debug("substituting '%s' for '%s' in commands matching '%s'" % (subst, pattern, globstr)) if not self.executed_commands: return 0 replacements = None try: for called in self.executed_commands: # was anything collected? if called['file'] is None: continue if called['binary'] == 'yes': self._log_warn("Cannot apply regex substitution to binary" " output: '%s'" % called['exe']) continue if fnmatch.fnmatch(called['cmd'], globstr): path = os.path.join(self.commons['cmddir'], called['file']) self._log_debug("applying substitution to '%s'" % path) readable = self.archive.open_file(path) result, replacements = re.subn( regexp, subst, readable.read()) if replacements: self.archive.add_string(result, path) except Exception as e: msg = "regex substitution failed for '%s' with: '%s'" self._log_error(msg % (called['exe'], e)) replacements = None return replacements
[docs] def do_file_private_sub(self, pathregex, desc=""): """Scrub certificate/key/etc information from files collected by sos. Files matching the provided pathregex are searched for content that resembles certificate, ssh keys, or similar information. Any matches are replaced with "-----SCRUBBED $desc" where `desc` is a description of the specific type of content being replaced, e.g. "-----SCRUBBED RSA PRIVATE KEY" so that support representatives can at least be informed of what type of content it was originally. :param pathregex: A string or regex of a filename to match against :type pathregex: ``str`` :param desc: A description of the replaced content :type desc: ``str`` """ self._log_debug("Scrubbing certs and keys for paths matching %s" % pathregex) match = re.compile(pathregex).match replace = "%s %s" % (_cert_replace, desc) if desc else _cert_replace file_list = [f for f in self.copied_files if match(f['srcpath'])] for i in file_list: path = i['dstpath'] if not path: continue self.do_file_sub(path, _certmatch, replace)
[docs] def do_file_sub(self, srcpath, regexp, subst): """Apply a regexp substitution to a file archived by sosreport. :param srcpath: Path in the archive where the file can be found :type srcpath: ``str`` :param regexp: A regex to match the contents of the file :type regexp: ``str`` or compiled ``re`` object :param subst: The substitution string to be used to replace matches within the file :type subst: ``str`` :returns: Number of replacements made :rtype: ``int`` """ try: path = self._get_dest_for_srcpath(srcpath) pattern = regexp.pattern if hasattr(regexp, "pattern") else regexp self._log_debug("substituting scrpath '%s'" % srcpath) self._log_debug("substituting '%s' for '%s' in '%s'" % (subst, pattern, path)) if not path: return 0 readable = self.archive.open_file(path) content = readable.read() if not isinstance(content, str): content = content.decode('utf8', 'ignore') result, replacements = re.subn(regexp, subst, content) if replacements: self.archive.add_string(result, srcpath) else: replacements = 0 except (OSError, IOError) as e: # if trying to regexp a nonexisting file, dont log it as an # error to stdout if e.errno == errno.ENOENT: msg = "file '%s' not collected, substitution skipped" self._log_debug(msg % path) else: msg = "regex substitution failed for '%s' with: '%s'" self._log_error(msg % (path, e)) replacements = 0 return replacements
[docs] def do_path_regex_sub(self, pathexp, regexp, subst): """Apply a regexp substituation to a set of files archived by sos. The set of files to be substituted is generated by matching collected file pathnames against `pathexp`. :param pathexp: A regex to match filenames within the archive :type pathexp: ``str`` or compiled ``re`` object :param regexp: A regex to match against the contents of each file :type regexp: ``str`` or compiled ``re`` object :param subst: The substituion string to be used to replace matches :type subst: ``str`` """ if not hasattr(pathexp, "match"): pathexp = re.compile(pathexp) match = pathexp.match file_list = [f for f in self.copied_files if match(f['srcpath'])] for file in file_list: self.do_file_sub(file['srcpath'], regexp, subst)
def do_regex_find_all(self, regex, fname): return regex_findall(regex, fname) def _copy_symlink(self, srcpath): # the target stored in the original symlink linkdest = os.readlink(srcpath) dest = os.path.join(os.path.dirname(srcpath), linkdest) # Absolute path to the link target. If SYSROOT != '/' this path # is relative to the host root file system. absdest = os.path.normpath(dest) # adjust the target used inside the report to always be relative if os.path.isabs(linkdest): # Canonicalize the link target path to avoid additional levels # of symbolic links (that would affect the path nesting level). realdir = os.path.realpath(os.path.dirname(srcpath)) reldest = os.path.relpath(linkdest, start=realdir) # trim leading /sysroot if self.use_sysroot(): reldest = reldest[len(os.sep + os.pardir):] self._log_debug("made link target '%s' relative as '%s'" % (linkdest, reldest)) else: reldest = linkdest self._log_debug("copying link '%s' pointing to '%s' with isdir=%s" % (srcpath, linkdest, os.path.isdir(absdest))) dstpath = self.strip_sysroot(srcpath) # use the relative target path in the tarball self.archive.add_link(reldest, dstpath) if os.path.isdir(absdest): self._log_debug("link '%s' is a directory, skipping..." % linkdest) return self.copied_files.append({'srcpath': srcpath, 'dstpath': dstpath, 'symlink': "yes", 'pointsto': linkdest}) # Check for indirect symlink loops by stat()ing the next step # in the link chain. try: os.stat(absdest) except OSError as e: if e.errno == 40: self._log_debug("link '%s' is part of a file system " "loop, skipping target..." % dstpath) return # copy the symlink target translating relative targets # to absolute paths to pass to _do_copy_path. self._log_debug("normalized link target '%s' as '%s'" % (linkdest, absdest)) # skip recursive copying of symlink pointing to itself. if (absdest != srcpath): self._do_copy_path(absdest) else: self._log_debug("link '%s' points to itself, skipping target..." % linkdest) def _copy_dir(self, srcpath): try: for name in os.listdir(srcpath): self._log_debug("recursively adding '%s' from '%s'" % (name, srcpath)) path = os.path.join(srcpath, name) self._do_copy_path(path) except OSError as e: if e.errno == errno.EPERM or errno.EACCES: msg = "Permission denied" self._log_warn("_copy_dir: '%s' %s" % (srcpath, msg)) return if e.errno == errno.ELOOP: msg = "Too many levels of symbolic links copying" self._log_error("_copy_dir: %s '%s'" % (msg, srcpath)) return raise def _get_dest_for_srcpath(self, srcpath): if self.use_sysroot(): srcpath = self.join_sysroot(srcpath) for copied in self.copied_files: if srcpath == copied["srcpath"]: return copied["dstpath"] return None def _is_forbidden_path(self, path): return _path_in_path_list(path, self.forbidden_paths) def _copy_node(self, path, st): dev_maj = os.major(st.st_rdev) dev_min = os.minor(st.st_rdev) mode = st.st_mode self.archive.add_node(path, mode, os.makedev(dev_maj, dev_min)) # Methods for copying files and shelling out def _do_copy_path(self, srcpath, dest=None): """Copy file or directory to the destination tree. If a directory, then everything below it is recursively copied. A list of copied files are saved for use later in preparing a report. """ if self._timeout_hit: return if self._is_forbidden_path(srcpath): self._log_debug("skipping forbidden path '%s'" % srcpath) return '' if not dest: dest = srcpath if self.use_sysroot(): dest = self.strip_sysroot(dest) try: st = os.lstat(srcpath) except (OSError, IOError): self._log_info("failed to stat '%s'" % srcpath) return if stat.S_ISLNK(st.st_mode): self._copy_symlink(srcpath) return else: if stat.S_ISDIR(st.st_mode) and os.access(srcpath, os.R_OK): # copy empty directory if not os.listdir(srcpath): self.archive.add_dir(dest) return self._copy_dir(srcpath) return # handle special nodes (block, char, fifo, socket) if not (stat.S_ISREG(st.st_mode) or stat.S_ISDIR(st.st_mode)): ntype = _node_type(st) self._log_debug("creating %s node at archive:'%s'" % (ntype, dest)) self._copy_node(dest, st) return # if we get here, it's definitely a regular file (not a symlink or dir) self._log_debug("copying path '%s' to archive:'%s'" % (srcpath, dest)) # if not readable(srcpath) if not st.st_mode & 0o444: # FIXME: reflect permissions in archive self.archive.add_string("", dest) else: self.archive.add_file(srcpath, dest) self.copied_files.append({ 'srcpath': srcpath, 'dstpath': dest, 'symlink': "no" })
[docs] def add_forbidden_path(self, forbidden): """Specify a path, or list of paths, to not copy, even if it's part of an ``add_copy_spec()`` call :param forbidden: A filepath to forbid collection from :type forbidden: ``str`` or a ``list`` of strings """ if isinstance(forbidden, str): forbidden = [forbidden] if self.use_sysroot(): forbidden = [self.join_sysroot(f) for f in forbidden] for forbid in forbidden: self._log_info("adding forbidden path '%s'" % forbid) for path in glob.glob(forbid): self.forbidden_paths.append(path)
[docs] def get_all_options(self): """return a list of all options selected""" return (self.opt_names, self.opt_parms)
[docs] def set_option(self, optionname, value): """Set the named option to value. Ensure the original type of the option value is preserved :param optioname: The name of the option to set :type optioname: ``str`` :param value: The value to set the option to :returns: ``True`` if the option is successfully set, else ``False`` :rtype: ``bool`` """ for name, parms in zip(self.opt_names, self.opt_parms): if name == optionname: # FIXME: ensure that the resulting type of the set option # matches that of the default value. This prevents a string # option from being coerced to int simply because it holds # a numeric value (e.g. a password). # See PR #1526 and Issue #1597 defaulttype = type(parms['enabled']) if defaulttype != type(value) and defaulttype != type(None): value = (defaulttype)(value) parms['enabled'] = value return True return False
[docs] def get_option(self, optionname, default=0): """Retrieve the value of the requested option, searching in order: parameters passed from the command line, set via `set_option()`, or the global_plugin_options dict. `optionname` may be iterable, in which case this function will return the first match. :param optionname: The name of the option to retrieve the value of :type optionname: ``str`` :param default: Optionally provide a default value to return if no option matching `optionname` is found. Default 0 :returns: The value of `optionname` if found, else `default` """ global_options = ( 'all_logs', 'allow_system_changes', 'log_size', 'plugin_timeout', 'since', 'verify' ) if optionname in global_options: return getattr(self.commons['cmdlineopts'], optionname) for name, parms in zip(self.opt_names, self.opt_parms): if name == optionname: val = parms['enabled'] if val is not None: return val return default
[docs] def get_option_as_list(self, optionname, delimiter=",", default=None): """Will try to return the option as a list separated by the delimiter. """ option = self.get_option(optionname) try: opt_list = [opt.strip() for opt in option.split(delimiter)] return list(filter(None, opt_list)) except Exception: return default
def _add_copy_paths(self, copy_paths): self.copy_paths.update(copy_paths)
[docs] def add_file_tags(self, tagdict): """Apply a tag to a file matching a given regex, for use when a file is copied by a more generic copyspec. :param tagdict: A dict containing the filepatterns to match and the tag(s) to apply to those files :type tagdict: ``dict`` `tagdict` takes the form `{file_pattern: tag}`, E.G. to match all bond devices from /proc/net/bonding with the tag `bond`, use `{'/proc/net/bonding/bond.*': ['bond']}` """ for fname in tagdict: if isinstance(tagdict[fname], str): tagdict[fname] = [tagdict[fname]] self.filetags.update(tagdict)
[docs] def get_tags_for_file(self, fname): """Get the tags that should be associated with a file matching a given regex :param fname: A regex for filenames to be matched against :type fname: ``str`` :returns: The tag(s) associated with `fname` :rtype: ``list`` of strings """ for key, val in self.filetags.items(): if re.match(key, fname): return val return []
[docs] def generate_copyspec_tags(self): """After file collections have completed, retroactively generate manifest entries to apply tags to files copied by generic copyspecs """ for file_regex in self.filetags: manifest_data = { 'specification': file_regex, 'files_copied': [], 'tags': self.filetags[file_regex] } matched_files = [] for cfile in self.copied_files: if re.match(file_regex, cfile['srcpath']): matched_files.append(cfile['dstpath']) if matched_files: manifest_data['files_copied'] = matched_files self.manifest.files.append(manifest_data)
[docs] def add_copy_spec(self, copyspecs, sizelimit=None, maxage=None, tailit=True, pred=None, tags=[]): """Add a file, directory, or regex matching filepaths to the archive :param copyspecs: A file, directory, or regex matching filepaths :type copyspecs: ``str`` or a ``list`` of strings :param sizelimit: Limit the total size of collections from `copyspecs` to this size in MB :type sizelimit: ``int`` :param maxage: Collect files with `mtime` not older than this many hours :type maxage: ``int`` :param tailit: Should a file that exceeds `sizelimit` be tail'ed to fit the remaining space to meet `sizelimit` :type tailit: ``bool`` :param pred: A predicate to gate if `copyspecs` should be collected :type pred: ``SoSPredicate`` :param tags: A tag or set of tags to add to the metadata information for this collection :type tags: ``str`` or a ``list`` of strings `copyspecs` will be expanded and/or globbed as appropriate. Specifying a directory here will cause the plugin to attempt to collect the entire directory, recursively. Note that `sizelimit` is applied to each `copyspec`, not each file individually. For example, a copyspec of ``['/etc/foo', '/etc/bar.conf']`` and a `sizelimit` of 25 means that sos will collect up to 25MB worth of files within `/etc/foo`, and will collect the last 25MB of `/etc/bar.conf`. """ since = None if self.get_option('since'): since = self.get_option('since') logarchive_pattern = re.compile(r'.*((\.(zip|gz|bz2|xz))|[-.][\d]+)$') configfile_pattern = re.compile(r"^%s/*" % self.join_sysroot("etc")) if not self.test_predicate(pred=pred): self._log_info("skipped copy spec '%s' due to predicate (%s)" % (copyspecs, self.get_predicate(pred=pred))) return if sizelimit is None: sizelimit = self.get_option("log_size") if self.get_option('all_logs'): sizelimit = None if sizelimit: sizelimit *= 1024 * 1024 # in MB if not copyspecs: return False if isinstance(copyspecs, str): copyspecs = [copyspecs] if isinstance(tags, str): tags = [tags] for copyspec in copyspecs: if not (copyspec and len(copyspec)): return False if self.use_sysroot(): copyspec = self.join_sysroot(copyspec) files = self._expand_copy_spec(copyspec) if len(files) == 0: continue def get_filename_tag(fname): """Generate a tag to add for a single file copyspec This tag will be set to the filename, minus any extensions except '.conf' which will be converted to '_conf' """ fname = fname.replace('-', '_') if fname.endswith('.conf'): return fname.replace('.', '_') return fname.split('.')[0] # Files hould be sorted in most-recently-modified order, so that # we collect the newest data first before reaching the limit. def getmtime(path): try: return os.path.getmtime(path) except OSError: return 0 def time_filter(path): """ When --since is passed, or maxage is coming from the plugin, we need to filter out older files """ # skip config files or not-logarchive files from the filter if ((logarchive_pattern.search(path) is None) or (configfile_pattern.search(path) is not None)): return True filetime = datetime.fromtimestamp(getmtime(path)) if ((since and filetime < since) or (maxage and (time()-filetime < maxage*3600))): return False return True _spec_tags = [] if len(files) == 1: _spec_tags = [get_filename_tag(files[0].split('/')[-1])] _spec_tags.extend(tags) if since or maxage: files = list(filter(lambda f: time_filter(f), files)) files.sort(key=getmtime, reverse=True) current_size = 0 limit_reached = False _manifest_files = [] for _file in files: if _file in self.copy_paths: self._log_debug("skipping redundant file '%s'" % _file) continue if self._is_forbidden_path(_file): self._log_debug("skipping forbidden path '%s'" % _file) continue if limit_reached: self._log_info("skipping '%s' over size limit" % _file) continue try: filestat = os.stat(_file) except OSError: self._log_info("failed to stat '%s'" % _file) continue current_size += filestat[stat.ST_SIZE] if sizelimit and current_size > sizelimit: limit_reached = True if tailit and not _file_is_compressed(_file): self._log_info("collecting tail of '%s' due to size " "limit" % _file) file_name = _file if file_name[0] == os.sep: file_name = file_name.lstrip(os.sep) strfile = ( file_name.replace(os.path.sep, ".") + ".tailed" ) add_size = (sizelimit + filestat[stat.ST_SIZE] - current_size) self.add_string_as_file(tail(_file, add_size), strfile) rel_path = os.path.relpath('/', os.path.dirname(_file)) link_path = os.path.join(rel_path, 'sos_strings', self.name(), strfile) self.archive.add_link(link_path, _file) _manifest_files.append(_file.lstrip('/')) else: self._log_info("skipping '%s' over size limit" % _file) else: # size limit not exceeded, copy the file _manifest_files.append(_file.lstrip('/')) self._add_copy_paths([_file]) # in the corner case we just reached the sizelimit, we # should collect the whole file and stop limit_reached = (current_size == sizelimit) if self.manifest: self.manifest.files.append({ 'specification': copyspec, 'files_copied': _manifest_files, 'tags': _spec_tags })
[docs] def add_blockdev_cmd(self, cmds, devices='block', timeout=300, sizelimit=None, chroot=True, runat=None, env=None, binary=False, prepend_path=None, whitelist=[], blacklist=[], tags=[]): """Run a command or list of commands against storage-related devices. Any commands specified by cmd will be iterated over the list of the specified devices. Commands passed to this should include a '%(dev)s' variable for substitution. :param cmds: The command(s) to run against the list of devices :type cmds: ``str`` or a ``list`` of strings :param devices: The device paths to run `cmd` against. If set to `block` or `fibre`, the commands will be run against the matching list of discovered devices :type devices: ``str`` or a ``list`` of device paths :param timeout: Timeout in seconds to allow each `cmd` to run :type timeout: ``int`` :param sizelimit: Maximum amount of output to collect, in MB :type sizelimit: ``int`` :param chroot: Should sos chroot the command(s) being run :type chroot: ``bool`` :param runat: Set the filesystem location to execute the command from :type runat: ``str`` :param env: Set environment variables for the command(s) being run :type env: ``dict`` :param binary: Is the output collected going to be binary data :type binary: ``bool`` :param prepend_path: The leading path for block device names :type prepend_path: ``str`` or ``None`` :param whitelist: Limit the devices the `cmds` will be run against to devices matching these item(s) :type whitelist: ``list`` of ``str`` :param blacklist: Do not run `cmds` against devices matching these item(s) :type blacklist: ``list`` of ``str`` """ _dev_tags = [] if isinstance(tags, str): tags = [tags] if devices == 'block': prepend_path = prepend_path or '/dev/' devices = self.devices['block'] _dev_tags.append('block') if devices == 'fibre': devices = self.devices['fibre'] _dev_tags.append('fibre') _dev_tags.extend(tags) self._add_device_cmd(cmds, devices, timeout=timeout, sizelimit=sizelimit, chroot=chroot, runat=runat, env=env, binary=binary, prepend_path=prepend_path, whitelist=whitelist, blacklist=blacklist, tags=_dev_tags)
def _add_device_cmd(self, cmds, devices, timeout=300, sizelimit=None, chroot=True, runat=None, env=None, binary=False, prepend_path=None, whitelist=[], blacklist=[], tags=[]): """Run a command against all specified devices on the system. """ if isinstance(cmds, str): cmds = [cmds] if isinstance(devices, str): devices = [devices] if isinstance(whitelist, str): whitelist = [whitelist] if isinstance(blacklist, str): blacklist = [blacklist] sizelimit = sizelimit or self.get_option('log_size') for cmd in cmds: for device in devices: _dev_ok = True _dev_tags = [device] _dev_tags.extend(tags) if whitelist: if not any(re.match(wl, device) for wl in whitelist): _dev_ok = False if blacklist: if any(re.match(blist, device) for blist in blacklist): _dev_ok = False if not _dev_ok: continue if prepend_path: device = os.path.join(prepend_path, device) _cmd = cmd % {'dev': device} self._add_cmd_output(cmd=_cmd, timeout=timeout, sizelimit=sizelimit, chroot=chroot, runat=runat, env=env, binary=binary, tags=_dev_tags) def _add_cmd_output(self, **kwargs): """Internal helper to add a single command to the collection list.""" pred = kwargs.pop('pred') if 'pred' in kwargs else None soscmd = SoSCommand(**kwargs) self._log_debug("packed command: " + soscmd.__str__()) if self.test_predicate(cmd=True, pred=pred): self.collect_cmds.append(soscmd) self._log_info("added cmd output '%s'" % soscmd.cmd) else: self.log_skipped_cmd(pred, soscmd.cmd, kmods=bool(pred.kmods), services=bool(pred.services), changes=soscmd.changes)
[docs] def add_cmd_output(self, cmds, suggest_filename=None, root_symlink=None, timeout=cmd_timeout, stderr=True, chroot=True, runat=None, env=None, binary=False, sizelimit=None, pred=None, subdir=None, changes=False, foreground=False, tags=[]): """Run a program or a list of programs and collect the output Output will be limited to `sizelimit`, collecting the last X amount of command output matching `sizelimit`. Unless `suggest_filename` is set, the file that the output is saved to will match the command as it was executed, and will be saved under `sos_commands/$plugin` :param cmds: The command(s) to execute :type cmds: ``str`` or a ``list`` of strings :param suggest_filename: Override the name of the file output is saved to within the archive :type suggest_filename: ``str`` :param root_symlink: If set, create a symlink with this name in the archive root :type root_symlink: ``str`` :param timeout: Timeout in seconds to allow each `cmd` to run for :type timeout: ``int`` :param stderr: Should stderr output be collected :type stderr: ``bool`` :param chroot: Should sos chroot the `cmds` being run :type chroot: ``bool`` :param runat: Run the `cmds` from this location in the filesystem :type runat: ``str`` :param env: Set environment variables for the `cmds` being run :type env: ``dict`` :param binary: Is the command expected to produce binary output :type binary: ``bool`` :param sizelimit: Maximum amount of output in MB to save :type sizelimit: ``int`` :param pred: A predicate to gate if `cmds` should be collected or not :type pred: ``SoSPredicate`` :param subdir: Save output to this subdirectory, within the plugin's directory under sos_commands :type subdir: ``str`` :param changes: Do `cmds` have the potential to change system state :type changes: ``int`` :param foreground: Should the `cmds` be run in the foreground, with an attached TTY :type foreground: ``bool`` :param tags: A tag or set of tags to add to the metadata entries for the `cmds` being run :type tags: ``str`` or a ``list`` of strings """ if isinstance(cmds, str): cmds = [cmds] if len(cmds) > 1 and (suggest_filename or root_symlink): self._log_warn("ambiguous filename or symlink for command list") if sizelimit is None: sizelimit = self.get_option("log_size") if pred is None: pred = self.get_predicate(cmd=True) for cmd in cmds: self._add_cmd_output(cmd=cmd, suggest_filename=suggest_filename, root_symlink=root_symlink, timeout=timeout, stderr=stderr, chroot=chroot, runat=runat, env=env, binary=binary, sizelimit=sizelimit, pred=pred, subdir=subdir, tags=tags, changes=changes, foreground=foreground)
[docs] def add_cmd_tags(self, tagdict): """Retroactively add tags to any commands that have been run by this plugin that match a given regex :param tagdict: A dict containing the command regex and associated tags :type tagdict: ``dict`` `tagdict` takes the form of {cmd_regex: tags}, for example to tag all commands starting with `foo` with the tag `bar`, use {'foo.*': ['bar']} """ for cmd in tagdict: if isinstance(tagdict[cmd], str): tagdict[cmd] = [tagdict[cmd]] self.cmdtags.update(tagdict)
[docs] def get_tags_for_cmd(self, cmd): """Get the tag(s) that should be associated with the given command :param cmd: The command that tags should be applied to :type cmd: ``str`` :returns: Any tags associated with the command :rtype: ``list`` """ for key, val in self.cmdtags.items(): if re.match(key, cmd): return val return []
[docs] def get_cmd_output_path(self, name=None, make=True): """Get the path where this plugin will save command output :param name: Optionally specify a filename to use as part of the command output path :type name: ``str`` or ``None`` :param make: Attempt to create the command output path :type make: ``bool`` :returns: The path where the plugin will write command output data within the archive :rtype: ``str`` """ cmd_output_path = os.path.join(self.archive.get_tmp_dir(), 'sos_commands', self.name()) if name: cmd_output_path = os.path.join(cmd_output_path, name) if make: os.makedirs(cmd_output_path) return cmd_output_path
[docs] def file_grep(self, regexp, *fnames): """Grep through file(s) for a specific string or regex :param regexp: The string or regex to search for :type regexp: ``str`` :param fnames: Paths to grep through :type fnames: ``str``, ``list`` of string, or open file objects :returns: Lines matching `regexp` :rtype: ``str`` """ return grep(regexp, *fnames)
def _mangle_command(self, exe): name_max = self.archive.name_max() return _mangle_command(exe, name_max) def _make_command_filename(self, exe, subdir=None): """The internal function to build up a filename based on a command.""" plugin_dir = self.name() if subdir: plugin_dir += "/%s" % subdir outdir = os.path.join(self.commons['cmddir'], plugin_dir) outfn = self._mangle_command(exe) # check for collisions if os.path.exists(os.path.join(self.archive.get_tmp_dir(), outdir, outfn)): inc = 1 name_max = self.archive.name_max() while True: suffix = ".%d" % inc newfn = outfn if name_max < len(newfn)+len(suffix): newfn = newfn[:(name_max-len(newfn)-len(suffix))] newfn = newfn + suffix if not os.path.exists(os.path.join(self.archive.get_tmp_dir(), outdir, newfn)): outfn = newfn break inc += 1 return os.path.join(outdir, outfn)
[docs] def add_env_var(self, name): """Add an environment variable to the list of to-be-collected env vars. Collected environment variables will be saved to an `environment` file in the archive root, and any variable specified for collection will be collected in lowercase, uppercase, and the form provided :param name: The name of the environment variable to collect :type name: ``str`` """ if not isinstance(name, list): name = [name] for env in name: # get both upper and lower cased vars since a common support issue # is setting the env vars to the wrong case, and if the plugin # adds a mixed case variable name, still get that as well self._env_vars.update([env, env.upper(), env.lower()])
[docs] def add_string_as_file(self, content, filename, pred=None): """Add a string to the archive as a file :param content: The string to write to the archive :type content: ``str`` :param filename: The name of the file to write `content` to :type filename: ``str`` :param pred: A predicate to gate if the string should be added to the archive or not :type pred: ``SoSPredicate`` """ # Generate summary string for logging summary = content.splitlines()[0] if content else '' if not isinstance(summary, str): summary = content.decode('utf8', 'ignore') if not self.test_predicate(cmd=False, pred=pred): self._log_info("skipped string ...'%s' due to predicate (%s)" % (summary, self.get_predicate(pred=pred))) return self.copy_strings.append((content, filename)) self._log_debug("added string ...'%s' as '%s'" % (summary, filename))
def _collect_cmd_output(self, cmd, suggest_filename=None, root_symlink=False, timeout=cmd_timeout, stderr=True, chroot=True, runat=None, env=None, binary=False, sizelimit=None, subdir=None, changes=False, foreground=False, tags=[]): """Execute a command and save the output to a file for inclusion in the report. Positional Arguments: :param cmd: The command to run Keyword Arguments: :param suggest_filename: Filename to use when writing to the archive :param root_symlink: Create a symlink in the archive root :param timeout: Time in seconds to allow a cmd to run :param stderr: Write stderr to stdout? :param chroot: Perform chroot before running cmd? :param runat: Run the command from this location, overriding chroot :param env: Dict of env vars to set for the cmd :param binary: Is the output in binary? :param sizelimit: Maximum size in MB of output to save :param subdir: Subdir in plugin directory to save to :param changes: Does this cmd potentially make a change on the system? :param tags: Add tags in the archive manifest :returns: dict containing status, output, and filename in the archive for the executed cmd """ if self._timeout_hit: return _tags = [] if isinstance(tags, str): tags = [tags] _tags.extend(tags) _tags.append(cmd.split(' ')[0]) _tags.extend(self.get_tags_for_cmd(cmd)) if chroot or self.commons['cmdlineopts'].chroot == 'always': root = self.sysroot else: root = None start = time() result = sos_get_command_output( cmd, timeout=timeout, stderr=stderr, chroot=root, chdir=runat, env=env, binary=binary, sizelimit=sizelimit, poller=self.check_timeout, foreground=foreground ) if result['status'] == 124: self._log_warn( "command '%s' timed out after %ds" % (cmd, timeout) ) manifest_cmd = { 'command': cmd.split(' ')[0], 'parameters': cmd.split(' ')[1:], 'exec': cmd, 'filepath': None, 'return_code': result['status'], 'run_time': time() - start, 'tags': _tags } # command not found or not runnable if result['status'] == 126 or result['status'] == 127: # automatically retry chroot'ed commands in the host namespace if root and root != '/': if self.commons['cmdlineopts'].chroot != 'always': self._log_info("command '%s' not found in %s - " "re-trying in host root" % (cmd.split()[0], root)) result = sos_get_command_output( cmd, timeout=timeout, chroot=False, chdir=runat, env=env, binary=binary, sizelimit=sizelimit, poller=self.check_timeout ) self._log_debug("could not run '%s': command not found" % cmd) # Exit here if the command was not found in the chroot check above # as otherwise we will create a blank file in the archive if result['status'] in [126, 127]: if self.manifest: self.manifest.commands.append(manifest_cmd) return result run_time = time() - start self._log_debug("collected output of '%s' in %s (changes=%s)" % (cmd.split()[0], run_time, changes)) if suggest_filename: outfn = self._make_command_filename(suggest_filename, subdir) else: outfn = self._make_command_filename(cmd, subdir) outfn_strip = outfn[len(self.commons['cmddir'])+1:] if binary: self.archive.add_binary(result['output'], outfn) else: self.archive.add_string(result['output'], outfn) if root_symlink: self.archive.add_link(outfn, root_symlink) # save info for later self.executed_commands.append({'cmd': cmd, 'file': outfn_strip, 'binary': 'yes' if binary else 'no'}) result['filename'] = ( os.path.join(self.archive.get_archive_path(), outfn) if outfn else '' ) if self.manifest: manifest_cmd['filepath'] = outfn manifest_cmd['run_time'] = run_time self.manifest.commands.append(manifest_cmd) return result
[docs] def collect_cmd_output(self, cmd, suggest_filename=None, root_symlink=False, timeout=cmd_timeout, stderr=True, chroot=True, runat=None, env=None, binary=False, sizelimit=None, pred=None, subdir=None, tags=[]): """Execute a command and save the output to a file for inclusion in the report, then return the results for further use by the plugin :param cmd: The command to run :type cmd: ``str`` :param suggest_filename: Filename to use when writing to the archive :param suggest_filename: ``str`` :param root_symlink: Create a symlink in the archive root :type root_symlink: ``bool`` :param timeout: Time in seconds to allow a cmd to run :type timeout: ``int`` :param stderr: Write stderr to stdout? :type stderr: ``bool`` :param chroot: Perform chroot before running cmd? :type chroot: ``bool`` :param runat: Run the command from this location, overriding chroot :type runat: ``str`` :param env: Environment vars to set for the cmd :type env: ``dict`` :param binary: Is the output in binary? :type binary: ``bool`` :param sizelimit: Maximum size in MB of output to save :type sizelimit: ``int`` :param subdir: Subdir in plugin directory to save to :type subdir: ``str`` :param changes: Does this cmd potentially make a change on the system? :type changes: ``bool`` :param tags: Add tags in the archive manifest :type tags: ``str`` or a ``list`` of strings :returns: `cmd` exit status, output, and the filepath within the archive output was saved to :rtype: ``dict`` """ if not self.test_predicate(cmd=True, pred=pred): self._log_info("skipped cmd output '%s' due to predicate (%s)" % (cmd, self.get_predicate(cmd=True, pred=pred))) return { 'status': None, # don't match on if result['status'] checks 'output': '', 'filename': '' } return self._collect_cmd_output( cmd, suggest_filename=suggest_filename, root_symlink=root_symlink, timeout=timeout, stderr=stderr, chroot=chroot, runat=runat, env=env, binary=binary, sizelimit=sizelimit, subdir=subdir, tags=tags )
[docs] def exec_cmd(self, cmd, timeout=cmd_timeout, stderr=True, chroot=True, runat=None, env=None, binary=False, pred=None, foreground=False, container=False): """Execute a command right now and return the output and status, but do not save the output within the archive. Use this method in a plugin's setup() if command output is needed to build subsequent commands added to a report via add_cmd_output(). :param cmd: The command to run :type cmd: ``str`` :param timeout: Time in seconds to allow a cmd to run :type timeout: ``int`` :param stderr: Write stderr to stdout? :type stderr: ``bool`` :param chroot: Perform chroot before running cmd? :type chroot: ``bool`` :param runat: Run the command from this location, overriding chroot :type runat: ``str`` :param env: Environment vars to set for the cmd :type env: ``dict`` :param binary: Is the output in binary? :type binary: ``bool`` :param pred: A predicate to gate execution of the `cmd` :type pred: ``SoSPredicate`` :param foreground: Run the `cmd` in the foreground with a TTY :type foreground: ``bool`` :param container: Execute this command in a container with this name :type container: ``str`` :returns: Command exit status and output :rtype: ``dict`` """ _default = {'status': None, 'output': ''} if not self.test_predicate(cmd=True, pred=pred): return _default if chroot or self.commons['cmdlineopts'].chroot == 'always': root = self.sysroot else: root = None if container: if self._get_container_runtime() is None: self._log_info("Cannot run cmd '%s' in container %s: no " "runtime detected on host." % (cmd, container)) return _default if self.container_exists(container): cmd = self.fmt_container_cmd(container, cmd) else: self._log_info("Cannot run cmd '%s' in container %s: no such " "container is running." % (cmd, container)) return sos_get_command_output(cmd, timeout=timeout, chroot=root, chdir=runat, binary=binary, env=env, foreground=foreground)
def _get_container_runtime(self, runtime=None): """Based on policy and request by the plugin, return a usable ContainerRuntime if one exists """ if runtime is None: if 'default' in self.policy.runtimes.keys(): return self.policy.runtimes['default'] else: for pol_runtime in list(self.policy.runtimes.keys()): if runtime == pol_runtime: return self.policy.runtimes[pol_runtime] return None
[docs] def container_exists(self, name): """If a container runtime is present, check to see if a container with a given name is currently running :param name: The name of the container to check presence of :type name: ``str`` :returns: ``True`` if `name` exists, else ``False`` :rtype: ``bool`` """ _runtime = self._get_container_runtime() if _runtime is not None: con = _runtime.get_container_by_name(name) return con is not None return False
[docs] def get_container_by_name(self, name): """Get the container ID for a specific container :param name: The name of the container :type name: ``str`` :returns: The ID of the container if it exists :rtype: ``str`` or ``None`` """ _runtime = self._get_container_runtime() if _runtime is not None: return _runtime.get_container_by_name(name) return None
[docs] def get_containers(self, runtime=None, get_all=False): """Return a list of all container IDs from the ``Policy`` ``ContainerRuntime`` If `runtime` is not provided, use the ``Policy`` default :param runtime: The container runtime to use, if not the default runtime detected and loaded by the ``Policy`` :type runtime: ``str`` :param get_all: Return all containers known to the `runtime`, even those that have terminated :type get_all: ``bool`` :returns: All container IDs found by the ``ContainerRuntime`` :rtype: ``list`` """ _runtime = self._get_container_runtime(runtime=runtime) if _runtime is not None: if get_all: return _runtime.get_containers(get_all=True) else: return _runtime.containers return []
[docs] def get_container_images(self, runtime=None): """Return a list of all image names from the Policy's ContainerRuntime If `runtime` is not provided, use the Policy default. If the specified `runtime` is not loaded, return empty. :param runtime: The container runtime to use, if not using the default runtime detected by the ``Policy`` :type runtime: ``str`` :returns: A list of container images known to the `runtime` :rtype: ``list`` """ _runtime = self._get_container_runtime(runtime=runtime) if _runtime is not None: return _runtime.images return []
[docs] def get_container_volumes(self, runtime=None): """Return a list of all volume names from the Policy's ContainerRuntime If `runtime` is not provided, use the Policy default. If the specified `runtime` is not loaded, return empty. :param runtime: The container runtime to use, if not using the default runtime detected by the ``Policy`` :type runtime: ``str`` :returns: A list of container volumes known to the `runtime` :rtype: ``list`` """ _runtime = self._get_container_runtime(runtime=runtime) if _runtime is not None: return _runtime.volumes return []
[docs] def get_container_logs(self, container, **kwargs): """Helper to get the ``logs`` output for a given container Supports passthru of add_cmd_output() options :param container: The name of the container to retrieve logs from :type container: ``str`` :param kwargs: Any kwargs supported by ``add_cmd_output()`` are supported here """ _runtime = self._get_container_runtime() if _runtime is not None: self.add_cmd_output(_runtime.get_logs_command(container), **kwargs)
[docs] def fmt_container_cmd(self, container, cmd): """Format a command to be executed by the loaded ``ContainerRuntime`` in a specified container :param container: The name of the container to execute the `cmd` in :type container: ``str`` :param cmd: The command to run within the container :type cmd: ``str`` :returns: The command to execute so that the specified `cmd` will run within the `container` and not on the host :rtype: ``str`` """ if self.container_exists(container): _runtime = self._get_container_runtime() return _runtime.fmt_container_cmd(container, cmd) return cmd
[docs] def is_module_loaded(self, module_name): """Determine whether specified module is loaded or not :param module_name: Name of kernel module to check for presence :type module_name: ``str`` :returns: ``True`` if the module is loaded, else ``False`` :rtype: ``bool`` """ return module_name in self.policy.kernel_mods
# For adding output
[docs] def add_alert(self, alertstring): """Add an alert to the collection of alerts for this plugin. These will be displayed in the report :param alertstring: The text to add as an alert :type alertstring: ``str`` """ self.alerts.append(alertstring)
[docs] def add_custom_text(self, text): """Append text to the custom text that is included in the report. This is freeform and can include html. :param text: The text to include in the report :type text: ``str`` """ self.custom_text += text
[docs] def add_service_status(self, services, **kwargs): """Collect service status information based on the ``InitSystem`` used :param services: Service name(s) to collect statuses for :type services: ``str`` or a ``list`` of strings :param kwargs: Optional arguments to pass to _add_cmd_output (timeout, predicate, suggest_filename,..) """ if isinstance(services, str): services = [services] query = self.policy.init_system.query_cmd if not query: # No policy defined InitSystem, cannot use add_service_status self._log_debug('Cannot add service output, policy does not define' ' an InitSystem to use') return for service in services: self._add_cmd_output(cmd="%s %s" % (query, service), **kwargs)
[docs] def add_journal(self, units=None, boot=None, since=None, until=None, lines=None, allfields=False, output=None, timeout=cmd_timeout, identifier=None, catalog=None, sizelimit=None, pred=None, tags=[]): """Collect journald logs from one of more units. :param units: Which journald units to collect :type units: ``str`` or a ``list`` of strings :param boot: A boot index using the journalctl syntax. The special values 'this' and 'last' are also accepted. :type boot: ``str`` :param since: Start time for journal messages :type since: ``str`` :param until: End time forjournal messages :type until: ``str`` :param lines: The maximum number of lines to be collected :type lines: ``int`` :param allfields: Include all journal fields regardless of size or non-printable characters :type allfields: ``bool`` :param output: Journalctl output control string, for example "verbose" :type output: ``str`` :param timeout: An optional timeout in seconds :type timeout: ``int`` :param identifier: An optional message identifier :type identifier: ``str`` :param catalog: Augment lines with descriptions from the system catalog :type catalog: ``bool`` :param sizelimit: Limit to the size of output returned in MB. Defaults to the value of --log-size. :type sizelimit: ``int`` """ journal_cmd = "journalctl --no-pager " unit_opt = " --unit %s" boot_opt = " --boot %s" since_opt = " --since '%s'" until_opt = " --until %s" lines_opt = " --lines %s" output_opt = " --output %s" identifier_opt = " --identifier %s" catalog_opt = " --catalog" journal_size = 100 all_logs = self.get_option("all_logs") log_size = sizelimit or self.get_option("log_size") log_size = max(log_size, journal_size) if not all_logs else 0 if isinstance(units, str): units = [units] if units: for unit in units: journal_cmd += unit_opt % unit if identifier: journal_cmd += identifier_opt % identifier if catalog: journal_cmd += catalog_opt if allfields: journal_cmd += " --all" if boot: if boot == "this": boot = "" if boot == "last": boot = "-1" journal_cmd += boot_opt % boot if since: journal_cmd += since_opt % since if until: journal_cmd += until_opt % until if lines: journal_cmd += lines_opt % lines if output: journal_cmd += output_opt % output self._log_debug("collecting journal: %s" % journal_cmd) self._add_cmd_output(cmd=journal_cmd, timeout=timeout, sizelimit=log_size, pred=pred, tags=tags)
def _expand_copy_spec(self, copyspec): def __expand(paths): found_paths = [] paths = glob.glob(paths) for path in paths: try: # avoid recursive symlink dirs if os.path.isfile(path) or os.path.islink(path): found_paths.append(path) elif os.path.isdir(path) and os.listdir(path): found_paths.extend(__expand(os.path.join(path, '*'))) else: found_paths.append(path) except PermissionError: # when running in LXD, we've seen os.access return True for # some /sys or /proc paths yet still get a PermissionError # when calling os.listdir(), so rather than rely on that, # just catch and ignore permissions errors resulting from # security modules like apparmor/selinux # Ref: https://github.com/lxc/lxd/issues/5688 pass return list(set(found_paths)) if (os.access(copyspec, os.R_OK) and os.path.isdir(copyspec) and os.listdir(copyspec)): # the directory exists and is non-empty, recurse through it copyspec = os.path.join(copyspec, '*') expanded = glob.glob(copyspec, recursive=True) recursed_files = [] for _path in expanded: try: if os.path.isdir(_path) and os.listdir(_path): # remove the top level dir to avoid duplicate attempts to # copy the dir and its contents expanded.remove(_path) recursed_files.extend(__expand(os.path.join(_path, '*'))) except PermissionError: # same as the above in __expand(), but this time remove the # path so we don't hit another PermissionError during the # actual copy expanded.remove(_path) expanded.extend(recursed_files) return list(set(expanded)) def _collect_copy_specs(self): for path in self.copy_paths: self._log_info("collecting path '%s'" % path) self._do_copy_path(path) self.generate_copyspec_tags() def _collect_cmds(self): for soscmd in self.collect_cmds: self._log_debug("unpacked command: " + soscmd.__str__()) self._log_info("collecting output of '%s'" % soscmd.cmd) self._collect_cmd_output(**soscmd.__dict__) def _collect_strings(self): for string, file_name in self.copy_strings: if self._timeout_hit: return content = '' if string: content = string.splitlines()[0] if not isinstance(content, str): content = content.decode('utf8', 'ignore') self._log_info("collecting string ...'%s' as '%s'" % (content, file_name)) try: self.archive.add_string(string, os.path.join('sos_strings', self.name(), file_name)) except Exception as e: self._log_debug("could not add string '%s': %s" % (file_name, e))
[docs] def collect(self): """Collect the data for a plugin.""" start = time() self._collect_copy_specs() self._collect_cmds() self._collect_strings() fields = (self.name(), time() - start) self._log_debug("collected plugin '%s' in %s" % fields)
[docs] def get_description(self): """This function will return the description for the plugin""" try: return self.short_desc except Exception: return "<no description available>"
[docs] def check_enabled(self): """This method will be used to verify that a plugin should execute given the condition of the underlying environment. The default implementation will return True if none of class.files, class.packages, nor class.commands is specified. If any of these is specified the plugin will check for the existence of any of the corresponding paths, packages or commands and return True if any are present. For SCLPlugin subclasses, it will check whether the plugin can be run for any of installed SCLs. If so, it will store names of these SCLs on the plugin class in addition to returning True. For plugins with more complex enablement checks this method may be overridden. :returns: ``True`` if the plugin should be run for this system, else ``False`` :rtype: ``bool`` """ # some files or packages have been specified for this package if any([self.files, self.packages, self.commands, self.kernel_mods, self.services, self.architectures]): if isinstance(self.files, str): self.files = [self.files] if isinstance(self.packages, str): self.packages = [self.packages] if isinstance(self.commands, str): self.commands = [self.commands] if isinstance(self.kernel_mods, str): self.kernel_mods = [self.kernel_mods] if isinstance(self.services, str): self.services = [self.services] if isinstance(self, SCLPlugin): # save SCLs that match files or packages type(self)._scls_matched = [] for scl in self._get_scls(): files = [f % {"scl_name": scl} for f in self.files] packages = [p % {"scl_name": scl} for p in self.packages] commands = [c % {"scl_name": scl} for c in self.commands] services = [s % {"scl_name": scl} for s in self.services] if self._check_plugin_triggers(files, packages, commands, services): type(self)._scls_matched.append(scl) return len(type(self)._scls_matched) > 0 return self._check_plugin_triggers(self.files, self.packages, self.commands, self.services) if isinstance(self, SCLPlugin): # if files and packages weren't specified, we take all SCLs type(self)._scls_matched = self._get_scls() return True
def _check_plugin_triggers(self, files, packages, commands, services): if not any([files, packages, commands, services]): # no checks beyond architecture restrictions return self.check_is_architecture() return ((any(os.path.exists(fname) for fname in files) or any(self.is_installed(pkg) for pkg in packages) or any(is_executable(cmd) for cmd in commands) or any(self.is_module_loaded(mod) for mod in self.kernel_mods) or any(self.is_service(svc) for svc in services)) and self.check_is_architecture())
[docs] def check_is_architecture(self): """Checks whether or not the system is running on an architecture that the plugin allows. If not architecture is set, assume plugin can run on all arches. :returns: ``True`` if the host's architecture allows the plugin to run, else ``False`` :rtype: ``bool`` """ if self.architectures is None: return True regex = '(?:%s)' % '|'.join(self.architectures) return re.match(regex, self.policy.get_arch())
[docs] def default_enabled(self): """This decides whether a plugin should be automatically loaded or only if manually specified in the command line.""" return True
[docs] def add_default_collections(self): """Based on the class attrs defined for plugin enablement, add a standardized set of collections before we call the plugin's own setup() method. """ # For any service used for enablement checks, collect its current # status if it exists for service in self.services: if self.is_service(service): self.add_service_status(service)
[docs] def setup(self): """Collect the list of files declared by the plugin. This method may be overridden to add further copy_specs, forbidden_paths, and external programs if required. """ self.add_copy_spec(list(self.files))
def setup_verify(self): if not hasattr(self, "verify_packages") or not self.verify_packages: if hasattr(self, "packages") and self.packages: # Limit automatic verification to only the named packages self.verify_packages = [p + "$" for p in self.packages] else: return pm = self.policy.package_manager verify_cmd = pm.build_verify_command(self.verify_packages) if verify_cmd: self.add_cmd_output(verify_cmd)
[docs] def postproc(self): """Perform any postprocessing. To be replaced by a plugin if required. """ pass
[docs] def check_process_by_name(self, process): """Checks if a named process is found in /proc/[0-9]*/cmdline. :param process: The name of the process :type process: ``str`` :returns: ``True`` if the process exists, else ``False`` :rtype: ``bool`` """ status = False cmd_line_glob = "/proc/[0-9]*/cmdline" try: cmd_line_paths = glob.glob(cmd_line_glob) for path in cmd_line_paths: f = open(path, 'r') cmd_line = f.read().strip() if process in cmd_line: status = True except IOError: return False return status
[docs] def get_process_pids(self, process): """Get a list of all PIDs that match a specified name :param process: The name of the process the get PIDs for :type process: ``str`` :returns: A list of PIDs :rtype: ``list`` """ pids = [] cmd_line_glob = "/proc/[0-9]*/cmdline" cmd_line_paths = glob.glob(cmd_line_glob) for path in cmd_line_paths: try: with open(path, 'r') as f: cmd_line = f.read().strip() if process in cmd_line: pids.append(path.split("/")[2]) except IOError: continue return pids
[docs]class RedHatPlugin(object): """Tagging class for Red Hat's Linux distributions""" pass
[docs]class SCLPlugin(RedHatPlugin): """Superclass for plugins operating on Software Collections (SCLs). Subclasses of this plugin class can specify class.files and class.packages using "%(scl_name)s" interpolation. The plugin invoking mechanism will try to match these against all found SCLs on the system. SCLs that do match class.files or class.packages are then accessible via self.scls_matched when the plugin is invoked. Additionally, this plugin class provides "add_cmd_output_scl" (run a command in context of given SCL), and "add_copy_spec_scl" and "add_copy_spec_limit_scl" (copy package from file system of given SCL). For example, you can implement a plugin that will list all global npm packages in every SCL that contains "npm" package: class SCLNpmPlugin(Plugin, SCLPlugin): packages = ("%(scl_name)s-npm",) def setup(self): for scl in self.scls_matched: self.add_cmd_output_scl(scl, "npm ls -g --json") """ @property def scls_matched(self): if not hasattr(type(self), '_scls_matched'): type(self)._scls_matched = [] return type(self)._scls_matched def _get_scls(self): output = sos_get_command_output("scl -l")["output"] return [scl.strip() for scl in output.splitlines()]
[docs] def convert_cmd_scl(self, scl, cmd): """wrapping command in "scl enable" call and adds proper PATH """ # load default SCL prefix to PATH prefix = self.policy.get_default_scl_prefix() # read prefix from /etc/scl/prefixes/${scl} and strip trailing '\n' try: prefix = open('/etc/scl/prefixes/%s' % scl, 'r').read()\ .rstrip('\n') except Exception as e: self._log_error("Failed to find prefix for SCL %s using %s: %s" % (scl, prefix, e)) # expand PATH by equivalent prefixes under the SCL tree path = os.environ["PATH"] for p in path.split(':'): path = '%s/%s%s:%s' % (prefix, scl, p, path) scl_cmd = "scl enable %s \"PATH=%s %s\"" % (scl, path, cmd) return scl_cmd
[docs] def add_cmd_output_scl(self, scl, cmds, **kwargs): """Same as add_cmd_output, except that it wraps command in "scl enable" call and sets proper PATH. """ if isinstance(cmds, str): cmds = [cmds] scl_cmds = [] for cmd in cmds: scl_cmds.append(self.convert_cmd_scl(scl, cmd)) self.add_cmd_output(scl_cmds, **kwargs)
# config files for Software Collections are under /etc/${prefix}/${scl} and # var files are under /var/${prefix}/${scl} where the ${prefix} is distro # specific path. So we need to insert the paths after the appropriate root # dir. def convert_copyspec_scl(self, scl, copyspec): scl_prefix = self.policy.get_default_scl_prefix() for rootdir in ['etc', 'var']: p = re.compile('^/%s/' % rootdir) copyspec = os.path.abspath(p.sub('/%s/%s/%s/' % (rootdir, scl_prefix, scl), copyspec)) return copyspec
[docs] def add_copy_spec_scl(self, scl, copyspecs): """Same as add_copy_spec, except that it prepends path to SCL root to "copyspecs". """ if isinstance(copyspecs, str): copyspecs = [copyspecs] scl_copyspecs = [] for copyspec in copyspecs: scl_copyspecs.append(self.convert_copyspec_scl(scl, copyspec)) self.add_copy_spec(scl_copyspecs)
[docs] def add_copy_spec_limit_scl(self, scl, copyspec, **kwargs): """Same as add_copy_spec_limit, except that it prepends path to SCL root to "copyspec". """ self.add_copy_spec_limit( self.convert_copyspec_scl(scl, copyspec), **kwargs )
[docs]class PowerKVMPlugin(RedHatPlugin): """Tagging class for IBM PowerKVM Linux""" pass
[docs]class ZKVMPlugin(RedHatPlugin): """Tagging class for IBM ZKVM Linux""" pass
[docs]class UbuntuPlugin(object): """Tagging class for Ubuntu Linux""" pass
[docs]class DebianPlugin(object): """Tagging class for Debian Linux""" pass
[docs]class SuSEPlugin(object): """Tagging class for SuSE Linux distributions""" pass
[docs]class CosPlugin(object): """Tagging class for Container-Optimized OS""" pass
[docs]class IndependentPlugin(object): """Tagging class for plugins that can run on any platform""" pass
[docs]class ExperimentalPlugin(object): """Tagging class that indicates that this plugin is experimental""" pass
[docs]def import_plugin(name, superclasses=None): """Import name as a module and return a list of all classes defined in that module. superclasses should be a tuple of valid superclasses to import, this defaults to (Plugin,). """ plugin_fqname = "sos.report.plugins.%s" % name if not superclasses: superclasses = (Plugin,) return import_module(plugin_fqname, superclasses)
# vim: set et ts=4 sw=4 :