sentence1
stringlengths
52
3.87M
sentence2
stringlengths
1
47.2k
label
stringclasses
1 value
def BVS(self, params): """ BVS label Branch to the instruction at label if the V flag is set """ label = self.get_one_parameter(self.ONE_PARAMETER, params) self.check_arguments(label_exists=(label,)) # BVS label def BVS_func(): if self.is_V_set(): self.register['PC'] = self.labels[label] return BVS_func
BVS label Branch to the instruction at label if the V flag is set
entailment
def get_queryset(self, request): """Shows one entry per distinct metric name""" queryset = super(MetricGroupAdmin, self).get_queryset(request) # poor-man's DISTINCT ON for Sqlite3 qs_values = queryset.values('id', 'name') # 2.7+ only :( # = {metric['name']: metric['id'] for metric in qs_values} distinct_names = {} for metric in qs_values: distinct_names[metric['name']] = metric['id'] queryset = self.model.objects.filter(id__in=distinct_names.values()) return queryset
Shows one entry per distinct metric name
entailment
def save_model(self, request, obj, form, change): """Updates all metrics with the same name""" like_metrics = self.model.objects.filter(name=obj.name) # 2.7+ only :( # = {key: form.cleaned_data[key] for key in form.changed_data} updates = {} for key in form.changed_data: updates[key] = form.cleaned_data[key] like_metrics.update(**updates)
Updates all metrics with the same name
entailment
def parse_lines(self, code): """ Return a list of the parsed code For each line, return a three-tuple containing: 1. The label 2. The instruction 3. Any arguments or parameters An element in the tuple may be None or '' if it did not find anything :param code: The code to parse :return: A list of tuples in the form of (label, instruction, parameters) """ remove_comments = re.compile(r'^([^;@\n]*);?.*$', re.MULTILINE) code = '\n'.join(remove_comments.findall(code)) # TODO can probably do this better # TODO labels with spaces between pipes is allowed `|label with space| INST OPER` parser = re.compile(r'^(\S*)?[\s]*(\S*)([^\n]*)$', re.MULTILINE) res = parser.findall(code) # Make all parsing of labels and instructions adhere to all uppercase res = [(label.upper(), instruction.upper(), parameters.strip()) for (label, instruction, parameters) in res] return res
Return a list of the parsed code For each line, return a three-tuple containing: 1. The label 2. The instruction 3. Any arguments or parameters An element in the tuple may be None or '' if it did not find anything :param code: The code to parse :return: A list of tuples in the form of (label, instruction, parameters)
entailment
def check_register(self, arg): """ Is the parameter a register in the form of 'R<d>', and if so is it within the bounds of registers defined Raises an exception if 1. The parameter is not in the form of 'R<d>' 2. <d> is outside the range of registers defined in the init value registers or _max_registers :param arg: The parameter to check :return: The number of the register """ self.check_parameter(arg) match = re.search(self.REGISTER_REGEX, arg) if match is None: raise iarm.exceptions.RuleError("Parameter {} is not a register".format(arg)) try: r_num = int(match.groups()[0]) except ValueError: r_num = int(match.groups()[0], 16) except TypeError: if arg in 'lr|LR': return 14 elif arg in 'sp|SP': return 13 elif arg in 'fp|FP': return 7 ## TODO this could be 7 or 11 depending on THUMB and ARM mode http://www.keil.com/support/man/docs/armcc/armcc_chr1359124947957.htm else: raise if r_num > self._max_registers: raise iarm.exceptions.RuleError( "Register {} is greater than defined registers of {}".format(arg, self._max_registers)) return r_num
Is the parameter a register in the form of 'R<d>', and if so is it within the bounds of registers defined Raises an exception if 1. The parameter is not in the form of 'R<d>' 2. <d> is outside the range of registers defined in the init value registers or _max_registers :param arg: The parameter to check :return: The number of the register
entailment
def check_immediate(self, arg): """ Is the parameter an immediate in the form of '#<d>', Raises an exception if 1. The parameter is not in the form of '#<d>' :param arg: The parameter to check :return: The value of the immediate """ self.check_parameter(arg) match = re.search(self.IMMEDIATE_REGEX, arg) if match is None: raise iarm.exceptions.RuleError("Parameter {} is not an immediate".format(arg)) return self.convert_to_integer(match.groups()[0])
Is the parameter an immediate in the form of '#<d>', Raises an exception if 1. The parameter is not in the form of '#<d>' :param arg: The parameter to check :return: The value of the immediate
entailment
def check_immediate_unsigned_value(self, arg, bit): """ Is the immediate within the unsigned value of 2**bit - 1 Raises an exception if 1. The immediate value is > 2**bit - 1 :param arg: The parameter to check :param bit: The number of bits to use in 2**bit :return: The value of the immediate """ i_num = self.check_immediate(arg) if (i_num > (2 ** bit - 1)) or (i_num < 0): raise iarm.exceptions.RuleError("Immediate {} is not an unsigned {} bit value".format(arg, bit)) return i_num
Is the immediate within the unsigned value of 2**bit - 1 Raises an exception if 1. The immediate value is > 2**bit - 1 :param arg: The parameter to check :param bit: The number of bits to use in 2**bit :return: The value of the immediate
entailment
def check_immediate_value(self, arg, _max, _min=0): """ Is the immediate within the range of [_min, _max] Raises an exception if 1. The immediate value is < _min or > _max :param arg: The parameter to check :param _max: The maximum value :param _min: The minimum value, optional, default is zero :return: The immediate value """ i_num = self.check_immediate(arg) if (i_num > _max) or (i_num < _min): raise iarm.exceptions.RuleError("Immediate {} is not within the range of [{}, {}]".format(arg, _min, _max)) return i_num
Is the immediate within the range of [_min, _max] Raises an exception if 1. The immediate value is < _min or > _max :param arg: The parameter to check :param _max: The maximum value :param _min: The minimum value, optional, default is zero :return: The immediate value
entailment
def rule_low_registers(self, arg): """Low registers are R0 - R7""" r_num = self.check_register(arg) if r_num > 7: raise iarm.exceptions.RuleError( "Register {} is not a low register".format(arg))
Low registers are R0 - R7
entailment
def get_parameters(self, regex_exp, parameters): """ Given a regex expression and the string with the paramers, either return a regex match object or raise an exception if the regex did not find a match :param regex_exp: :param parameters: :return: """ # TODO find a better way to do the equate replacement for rep in self.equates: parameters = parameters.replace(rep, str(self.equates[rep])) match = re.match(regex_exp, parameters) if not match: raise iarm.exceptions.ParsingError("Parameters are None, did you miss a comma?") return match.groups()
Given a regex expression and the string with the paramers, either return a regex match object or raise an exception if the regex did not find a match :param regex_exp: :param parameters: :return:
entailment
def get_one_parameter(self, regex_exp, parameters): """ Get three parameters from a given regex expression Raise an exception if more than three were found :param regex_exp: :param parameters: :return: """ Rx, other = self.get_parameters(regex_exp, parameters) if other is not None and other.strip(): raise iarm.exceptions.ParsingError("Extra arguments found: {}".format(other)) return Rx.upper()
Get three parameters from a given regex expression Raise an exception if more than three were found :param regex_exp: :param parameters: :return:
entailment
def get_two_parameters(self, regex_exp, parameters): """ Get two parameters from a given regex expression Raise an exception if more than two were found :param regex_exp: :param parameters: :return: """ Rx, Ry, other = self.get_parameters(regex_exp, parameters) if other is not None and other.strip(): raise iarm.exceptions.ParsingError("Extra arguments found: {}".format(other)) if Rx and Ry: return Rx.upper(), Ry.upper() elif not Rx: raise iarm.exceptions.ParsingError("Missing first positional argument") else: raise iarm.exceptions.ParsingError("Missing second positional argument")
Get two parameters from a given regex expression Raise an exception if more than two were found :param regex_exp: :param parameters: :return:
entailment
def get_three_parameters(self, regex_exp, parameters): """ Get three parameters from a given regex expression Raise an exception if more than three were found :param regex_exp: :param parameters: :return: """ Rx, Ry, Rz, other = self.get_parameters(regex_exp, parameters) if other is not None and other.strip(): raise iarm.exceptions.ParsingError("Extra arguments found: {}".format(other)) return Rx.upper(), Ry.upper(), Rz.upper()
Get three parameters from a given regex expression Raise an exception if more than three were found :param regex_exp: :param parameters: :return:
entailment
def set_APSR_flag_to_value(self, flag, value): """ Set or clear flag in ASPR :param flag: The flag to set :param value: If value evaulates to true, it is set, cleared otherwise :return: """ if flag == 'N': bit = 31 elif flag == 'Z': bit = 30 elif flag == 'C': bit = 29 elif flag == 'V': bit = 28 else: raise AttributeError("Flag {} does not exist in the APSR".format(flag)) if value: self.register['APSR'] |= (1 << bit) else: self.register['APSR'] -= (1 << bit) if (self.register['APSR'] & (1 << bit)) else 0
Set or clear flag in ASPR :param flag: The flag to set :param value: If value evaulates to true, it is set, cleared otherwise :return:
entailment
def rule_special_registers(self, arg): """Raises an exception if the register is not a special register""" # TODO is PSR supposed to be here? special_registers = "PSR APSR IPSR EPSR PRIMASK FAULTMASK BASEPRI CONTROL" if arg not in special_registers.split(): raise iarm.exceptions.RuleError("{} is not a special register; Must be [{}]".format(arg, special_registers))
Raises an exception if the register is not a special register
entailment
def set_C_flag(self, oper_1, oper_2, result, type): """ Set C flag C flag is set if the unsigned number overflows This condition is obtained if: 1. In addition, the result is smaller than either of the operands 2. In subtraction, if the second operand is larger than the first This should not be used for shifting as each shift will need to set the C flag differently """ # TODO is this correct? if type == 'add': if result < oper_1: self.set_APSR_flag_to_value('C', 1) else: self.set_APSR_flag_to_value('C', 0) elif type == 'sub': if oper_1 < oper_2: # If there was a borrow, then set to zero self.set_APSR_flag_to_value('C', 0) else: self.set_APSR_flag_to_value('C', 1) elif type == 'shift-left': if (oper_2 > 0) and (oper_2 < (self._bit_width - 1)): self.set_APSR_flag_to_value('C', oper_1 & (1 << (self._bit_width - oper_2))) else: self.set_APSR_flag_to_value('C', 0) else: raise iarm.exceptions.BrainFart("_type is not 'add' or 'sub'")
Set C flag C flag is set if the unsigned number overflows This condition is obtained if: 1. In addition, the result is smaller than either of the operands 2. In subtraction, if the second operand is larger than the first This should not be used for shifting as each shift will need to set the C flag differently
entailment
def version(): """Output version of gcdt tools and plugins.""" log.info('gcdt version %s' % __version__) tools = get_plugin_versions('gcdttool10') if tools: log.info('gcdt tools:') for p, v in tools.items(): log.info(' * %s version %s' % (p, v)) log.info('gcdt plugins:') for p, v in get_plugin_versions().items(): log.info(' * %s version %s' % (p, v)) generators = get_plugin_versions('gcdtgen10') if generators: log.info('gcdt scaffolding generators:') for p, v in generators.items(): log.info(' * %s version %s' % (p, v))
Output version of gcdt tools and plugins.
entailment
def retries(max_tries, delay=1, backoff=2, exceptions=(Exception,), hook=None): """Function decorator implementing retrying logic. delay: Sleep this many seconds * backoff * try number after failure backoff: Multiply delay by this factor after each failure exceptions: A tuple of exception classes; default (Exception,) hook: A function with the signature: (tries_remaining, exception, mydelay) """ """ def example_hook(tries_remaining, exception, delay): '''Example exception handler; prints a warning to stderr. tries_remaining: The number of tries remaining. exception: The exception instance which was raised. ''' print >> sys.stderr, "Caught '%s', %d tries remaining, sleeping for %s seconds" % (exception, tries_remaining, delay) The decorator will call the function up to max_tries times if it raises an exception. By default it catches instances of the Exception class and subclasses. This will recover after all but the most fatal errors. You may specify a custom tuple of exception classes with the 'exceptions' argument; the function will only be retried if it raises one of the specified exceptions. Additionally you may specify a hook function which will be called prior to retrying with the number of remaining tries and the exception instance; see given example. This is primarily intended to give the opportunity to log the failure. Hook is not called after failure if no retries remain. """ def dec(func): def f2(*args, **kwargs): mydelay = delay #tries = range(max_tries) #tries.reverse() tries = range(max_tries-1, -1, -1) for tries_remaining in tries: try: return func(*args, **kwargs) except GracefulExit: raise except exceptions as e: if tries_remaining > 0: if hook is not None: hook(tries_remaining, e, mydelay) sleep(mydelay) mydelay *= backoff else: raise return f2 return dec
Function decorator implementing retrying logic. delay: Sleep this many seconds * backoff * try number after failure backoff: Multiply delay by this factor after each failure exceptions: A tuple of exception classes; default (Exception,) hook: A function with the signature: (tries_remaining, exception, mydelay)
entailment
def get_env(): """ Read environment from ENV and mangle it to a (lower case) representation Note: gcdt.utils get_env() is used in many cloudformation.py templates :return: Environment as lower case string (or None if not matched) """ env = os.getenv('ENV', os.getenv('env', None)) if env: env = env.lower() return env
Read environment from ENV and mangle it to a (lower case) representation Note: gcdt.utils get_env() is used in many cloudformation.py templates :return: Environment as lower case string (or None if not matched)
entailment
def get_context(awsclient, env, tool, command, arguments=None): """This assembles the tool context. Private members are preceded by a '_'. :param tool: :param command: :return: dictionary containing the gcdt tool context """ # TODO: elapsed, artifact(stack, depl-grp, lambda, api) if arguments is None: arguments = {} context = { '_awsclient': awsclient, 'env': env, 'tool': tool, 'command': command, '_arguments': arguments, # TODO clean up arguments -> args 'version': __version__, 'user': _get_user(), 'plugins': get_plugin_versions().keys() } return context
This assembles the tool context. Private members are preceded by a '_'. :param tool: :param command: :return: dictionary containing the gcdt tool context
entailment
def get_command(arguments): """Extract the first argument from arguments parsed by docopt. :param arguments parsed by docopt: :return: command """ return [k for k, v in arguments.items() if not k.startswith('-') and v is True][0]
Extract the first argument from arguments parsed by docopt. :param arguments parsed by docopt: :return: command
entailment
def check_gcdt_update(): """Check whether a newer gcdt is available and output a warning. """ try: inst_version, latest_version = get_package_versions('gcdt') if inst_version < latest_version: log.warn('Please consider an update to gcdt version: %s' % latest_version) except GracefulExit: raise except Exception: log.warn('PyPi appears to be down - we currently can\'t check for newer gcdt versions')
Check whether a newer gcdt is available and output a warning.
entailment
def dict_selective_merge(a, b, selection, path=None): """Conditionally merges b into a if b's keys are contained in selection :param a: :param b: :param selection: limit merge to these top-level keys :param path: :return: """ if path is None: path = [] for key in b: if key in selection: if key in a: if isinstance(a[key], dict) and isinstance(b[key], dict): dict_selective_merge(a[key], b[key], b[key].keys(), path + [str(key)]) elif a[key] != b[key]: # update the value a[key] = b[key] else: a[key] = b[key] return a
Conditionally merges b into a if b's keys are contained in selection :param a: :param b: :param selection: limit merge to these top-level keys :param path: :return:
entailment
def dict_merge(a, b, path=None): """merges b into a""" return dict_selective_merge(a, b, b.keys(), path)
merges b into a
entailment
def are_credentials_still_valid(awsclient): """Check whether the credentials have expired. :param awsclient: :return: exit_code """ client = awsclient.get_client('lambda') try: client.list_functions() except GracefulExit: raise except Exception as e: log.debug(e) log.error(e) return 1 return 0
Check whether the credentials have expired. :param awsclient: :return: exit_code
entailment
def flatten(lis): """Given a list, possibly nested to any level, return it flattened.""" new_lis = [] for item in lis: if isinstance(item, collections.Sequence) and not isinstance(item, basestring): new_lis.extend(flatten(item)) else: new_lis.append(item) return new_lis
Given a list, possibly nested to any level, return it flattened.
entailment
def signal_handler(signum, frame): """ handle signals. example: 'signal.signal(signal.SIGTERM, signal_handler)' """ # signals are CONSTANTS so there is no mapping from signum to description # so please add to the mapping in case you use more signals! description = '%d' % signum if signum == 2: description = 'SIGINT' elif signum == 15: description = 'SIGTERM' raise GracefulExit(description)
handle signals. example: 'signal.signal(signal.SIGTERM, signal_handler)'
entailment
def json2table(json): """This does format a dictionary into a table. Note this expects a dictionary (not a json string!) :param json: :return: """ filter_terms = ['ResponseMetadata'] table = [] try: for k in filter(lambda k: k not in filter_terms, json.keys()): table.append([k.encode('ascii', 'ignore'), str(json[k]).encode('ascii', 'ignore')]) return tabulate(table, tablefmt='fancy_grid') except GracefulExit: raise except Exception as e: log.error(e) return json
This does format a dictionary into a table. Note this expects a dictionary (not a json string!) :param json: :return:
entailment
def random_string(length=6): """Create a random 6 character string. note: in case you use this function in a test during test together with an awsclient then this function is altered so you get reproducible results that will work with your recorded placebo json files (see helpers_aws.py). """ return ''.join([random.choice(string.ascii_lowercase) for i in range(length)])
Create a random 6 character string. note: in case you use this function in a test during test together with an awsclient then this function is altered so you get reproducible results that will work with your recorded placebo json files (see helpers_aws.py).
entailment
def all_pages(method, request, accessor, cond=None): """Helper to process all pages using botocore service methods (exhausts NextToken). note: `cond` is optional... you can use it to make filtering more explicit if you like. Alternatively you can do the filtering in the `accessor` which is perfectly fine, too Note: lambda uses a slightly different mechanism so there is a specific version in ramuda_utils. :param method: service method :param request: request dictionary for service call :param accessor: function to extract data from each response :param cond: filter function to return True / False based on a response :return: list of collected resources """ if cond is None: cond = lambda x: True result = [] next_token = None while True: if next_token: request['nextToken'] = next_token response = method(**request) if cond(response): data = accessor(response) if data: if isinstance(data, list): result.extend(data) else: result.append(data) if 'nextToken' not in response: break next_token = response['nextToken'] return result
Helper to process all pages using botocore service methods (exhausts NextToken). note: `cond` is optional... you can use it to make filtering more explicit if you like. Alternatively you can do the filtering in the `accessor` which is perfectly fine, too Note: lambda uses a slightly different mechanism so there is a specific version in ramuda_utils. :param method: service method :param request: request dictionary for service call :param accessor: function to extract data from each response :param cond: filter function to return True / False based on a response :return: list of collected resources
entailment
def get_invoke(command): """ Get the Cmd main method from the click command :param command: The click Command object :return: the do_* method for Cmd :rtype: function """ assert isinstance(command, click.Command) def invoke_(self, arg): # pylint: disable=unused-argument try: command.main(args=shlex.split(arg), prog_name=command.name, standalone_mode=False, parent=self.ctx) except click.ClickException as e: # Show the error message e.show() except click.Abort: # We got an EOF or Keyboard interrupt. Just silence it pass except SystemExit: # Catch this an return the code instead. All of click's help commands do a sys.exit(), # and that's not ideal when running in a shell. pass except Exception as e: traceback.print_exception(type(e), e, None) logger.warning(traceback.format_exc()) # Always return False so the shell doesn't exit return False invoke_ = update_wrapper(invoke_, command.callback) invoke_.__name__ = 'do_%s' % command.name return invoke_
Get the Cmd main method from the click command :param command: The click Command object :return: the do_* method for Cmd :rtype: function
entailment
def get_help(command): """ Get the Cmd help function from the click command :param command: The click Command object :return: the help_* method for Cmd :rtype: function """ assert isinstance(command, click.Command) def help_(self): # pylint: disable=unused-argument extra = {} for key, value in command.context_settings.items(): if key not in extra: extra[key] = value # Print click's help message with click.Context(command, info_name=command.name, parent=self.ctx, **extra) as ctx: click.echo(ctx.get_help(), color=ctx.color) help_.__name__ = 'help_%s' % command.name return help_
Get the Cmd help function from the click command :param command: The click Command object :return: the help_* method for Cmd :rtype: function
entailment
def get_complete(command): """ Get the Cmd complete function for the click command :param command: The click Command object :return: the complete_* method for Cmd :rtype: function """ assert isinstance(command, click.Command) def complete_(self, text, line, begidx, endidx): # pylint: disable=unused-argument # Parse the args args = shlex.split(line[:begidx]) # Strip of the first item which is the name of the command args = args[1:] # Then pass them on to the get_choices method that click uses for completion return list(get_choices(command, command.name, args, text)) complete_.__name__ = 'complete_%s' % command.name return complete_
Get the Cmd complete function for the click command :param command: The click Command object :return: the complete_* method for Cmd :rtype: function
entailment
def load_template(): """Bail out if template is not found. """ cloudformation, found = load_cloudformation_template() if not found: print(colored.red('could not load cloudformation.py, bailing out...')) sys.exit(1) return cloudformation
Bail out if template is not found.
entailment
def _import_public_names(module): "Import public names from module into this module, like import *" self = sys.modules[__name__] for name in module.__all__: if hasattr(self, name): # don't overwrite existing names continue setattr(self, name, getattr(module, name))
Import public names from module into this module, like import *
entailment
def create_argument_parser(): """Creates the argument parser for haas. """ parser = argparse.ArgumentParser(prog='haas') parser.add_argument('--version', action='version', version='%(prog)s {0}'.format(haas.__version__)) verbosity = parser.add_mutually_exclusive_group() verbosity.add_argument('-v', '--verbose', action='store_const', default=1, dest='verbosity', const=2, help='Verbose output') verbosity.add_argument('-q', '--quiet', action='store_const', const=0, dest='verbosity', help='Quiet output') parser.add_argument('-f', '--failfast', action='store_true', default=False, help='Stop on first fail or error') parser.add_argument('-c', '--catch', dest='catch_interrupt', action='store_true', default=False, help=('(Ignored) Catch ctrl-C and display results so ' 'far')) parser.add_argument('-b', '--buffer', action='store_true', default=False, help='Buffer stdout and stderr during tests') parser.add_argument( 'start', nargs='*', default=[os.getcwd()], help=('One or more directories or dotted package/module names from ' 'which to start searching for tests')) parser.add_argument('-p', '--pattern', default='test*.py', help="Pattern to match tests ('test*.py' default)") parser.add_argument('-t', '--top-level-directory', default=None, help=('Top level directory of project (defaults to ' 'start directory)')) _add_log_level_option(parser) return parser
Creates the argument parser for haas.
entailment
def run(self, plugin_manager=None): """Run the haas test runner. This will load and configure the selected plugins, set up the environment and begin test discovery, loading and running. Parameters ---------- plugin_manager : haas.plugin_manager.PluginManager [Optional] Override the use of the default plugin manager. """ if plugin_manager is None: plugin_manager = PluginManager() plugin_manager.add_plugin_arguments(self.parser) args = self.parser.parse_args(self.argv[1:]) environment_plugins = plugin_manager.get_enabled_hook_plugins( plugin_manager.ENVIRONMENT_HOOK, args) runner = plugin_manager.get_driver( plugin_manager.TEST_RUNNER, args) with PluginContext(environment_plugins): loader = Loader() discoverer = plugin_manager.get_driver( plugin_manager.TEST_DISCOVERY, args, loader=loader) suites = [ discoverer.discover( start=start, top_level_directory=args.top_level_directory, pattern=args.pattern, ) for start in args.start ] if len(suites) == 1: suite = suites[0] else: suite = loader.create_suite(suites) test_count = suite.countTestCases() result_handlers = plugin_manager.get_enabled_hook_plugins( plugin_manager.RESULT_HANDLERS, args, test_count=test_count) result_collector = ResultCollector( buffer=args.buffer, failfast=args.failfast) for result_handler in result_handlers: result_collector.add_result_handler(result_handler) result = runner.run(result_collector, suite) return not result.wasSuccessful()
Run the haas test runner. This will load and configure the selected plugins, set up the environment and begin test discovery, loading and running. Parameters ---------- plugin_manager : haas.plugin_manager.PluginManager [Optional] Override the use of the default plugin manager.
entailment
def encrypt_password(**kwargs): """ Crypt :new_value: if it's not crypted yet. """ new_value = kwargs['new_value'] field = kwargs['field'] min_length = field.params['min_length'] if len(new_value) < min_length: raise ValueError( '`{}`: Value length must be more than {}'.format( field.name, field.params['min_length'])) if new_value and not crypt.match(new_value): new_value = str(crypt.encode(new_value)) return new_value
Crypt :new_value: if it's not crypted yet.
entailment
def create_apikey_model(user_model): """ Generate ApiKey model class and connect it with :user_model:. ApiKey is generated with relationship to user model class :user_model: as a One-to-One relationship with a backreference. ApiKey is set up to be auto-generated when a new :user_model: is created. Returns ApiKey document class. If ApiKey is already defined, it is not generated. Arguments: :user_model: Class that represents user model for which api keys will be generated and with which ApiKey will have relationship. """ try: return engine.get_document_cls('ApiKey') except ValueError: pass fk_kwargs = { 'ref_column': None, } if hasattr(user_model, '__tablename__'): fk_kwargs['ref_column'] = '.'.join([ user_model.__tablename__, user_model.pk_field()]) fk_kwargs['ref_column_type'] = user_model.pk_field_type() class ApiKey(engine.BaseDocument): __tablename__ = 'nefertari_apikey' id = engine.IdField(primary_key=True) token = engine.StringField(default=create_apikey_token) user = engine.Relationship( document=user_model.__name__, uselist=False, backref_name='api_key', backref_uselist=False) user_id = engine.ForeignKeyField( ref_document=user_model.__name__, **fk_kwargs) def reset_token(self): self.update({'token': create_apikey_token()}) return self.token # Setup ApiKey autogeneration on :user_model: creation ApiKey.autogenerate_for(user_model, 'user') return ApiKey
Generate ApiKey model class and connect it with :user_model:. ApiKey is generated with relationship to user model class :user_model: as a One-to-One relationship with a backreference. ApiKey is set up to be auto-generated when a new :user_model: is created. Returns ApiKey document class. If ApiKey is already defined, it is not generated. Arguments: :user_model: Class that represents user model for which api keys will be generated and with which ApiKey will have relationship.
entailment
def cache_request_user(user_cls, request, user_id): """ Helper function to cache currently logged in user. User is cached at `request._user`. Caching happens only only if user is not already cached or if cached user's pk does not match `user_id`. :param user_cls: User model class to use for user lookup. :param request: Pyramid Request instance. :user_id: Current user primary key field value. """ pk_field = user_cls.pk_field() user = getattr(request, '_user', None) if user is None or getattr(user, pk_field, None) != user_id: request._user = user_cls.get_item(**{pk_field: user_id})
Helper function to cache currently logged in user. User is cached at `request._user`. Caching happens only only if user is not already cached or if cached user's pk does not match `user_id`. :param user_cls: User model class to use for user lookup. :param request: Pyramid Request instance. :user_id: Current user primary key field value.
entailment
def get_token_credentials(cls, username, request): """ Get api token for user with username of :username: Used by Token-based auth as `credentials_callback` kwarg. """ try: user = cls.get_item(username=username) except Exception as ex: log.error(str(ex)) forget(request) else: if user: return user.api_key.token
Get api token for user with username of :username: Used by Token-based auth as `credentials_callback` kwarg.
entailment
def get_groups_by_token(cls, username, token, request): """ Get user's groups if user with :username: exists and their api key token equals :token: Used by Token-based authentication as `check` kwarg. """ try: user = cls.get_item(username=username) except Exception as ex: log.error(str(ex)) forget(request) return else: if user and user.api_key.token == token: return ['g:%s' % g for g in user.groups]
Get user's groups if user with :username: exists and their api key token equals :token: Used by Token-based authentication as `check` kwarg.
entailment
def authenticate_by_password(cls, params): """ Authenticate user with login and password from :params: Used both by Token and Ticket-based auths (called from views). """ def verify_password(user, password): return crypt.check(user.password, password) success = False user = None login = params['login'].lower().strip() key = 'email' if '@' in login else 'username' try: user = cls.get_item(**{key: login}) except Exception as ex: log.error(str(ex)) if user: password = params.get('password', None) success = (password and verify_password(user, password)) return success, user
Authenticate user with login and password from :params: Used both by Token and Ticket-based auths (called from views).
entailment
def get_groups_by_userid(cls, userid, request): """ Return group identifiers of user with id :userid: Used by Ticket-based auth as `callback` kwarg. """ try: cache_request_user(cls, request, userid) except Exception as ex: log.error(str(ex)) forget(request) else: if request._user: return ['g:%s' % g for g in request._user.groups]
Return group identifiers of user with id :userid: Used by Ticket-based auth as `callback` kwarg.
entailment
def create_account(cls, params): """ Create auth user instance with data from :params:. Used by both Token and Ticket-based auths to register a user ( called from views). """ user_params = dictset(params).subset( ['username', 'email', 'password']) try: return cls.get_or_create( email=user_params['email'], defaults=user_params) except JHTTPBadRequest: raise JHTTPBadRequest('Failed to create account.')
Create auth user instance with data from :params:. Used by both Token and Ticket-based auths to register a user ( called from views).
entailment
def get_authuser_by_userid(cls, request): """ Get user by ID. Used by Ticket-based auth. Is added as request method to populate `request.user`. """ userid = authenticated_userid(request) if userid: cache_request_user(cls, request, userid) return request._user
Get user by ID. Used by Ticket-based auth. Is added as request method to populate `request.user`.
entailment
def get_authuser_by_name(cls, request): """ Get user by username Used by Token-based auth. Is added as request method to populate `request.user`. """ username = authenticated_userid(request) if username: return cls.get_item(username=username)
Get user by username Used by Token-based auth. Is added as request method to populate `request.user`.
entailment
def walk(self): """ Walk the directory like os.path (yields a 3-tuple (dirpath, dirnames, filenames) except it exclude all files/directories on the fly. """ for root, dirs, files in os.walk(self.path, topdown=True): # TODO relative walk, recursive call if root excluder found??? #root_excluder = get_root_excluder(root) ndirs = [] # First we exclude directories for d in list(dirs): if self.is_excluded(os.path.join(root, d)): dirs.remove(d) elif not os.path.islink(os.path.join(root, d)): ndirs.append(d) nfiles = [] for fpath in (os.path.join(root, f) for f in files): if not self.is_excluded(fpath) and not os.path.islink(fpath): nfiles.append(os.path.relpath(fpath, root)) yield root, ndirs, nfiles
Walk the directory like os.path (yields a 3-tuple (dirpath, dirnames, filenames) except it exclude all files/directories on the fly.
entailment
def includeme(config): """ Set up event subscribers. """ from .models import ( AuthUserMixin, random_uuid, lower_strip, encrypt_password, ) add_proc = config.add_field_processors add_proc( [random_uuid, lower_strip], model=AuthUserMixin, field='username') add_proc([lower_strip], model=AuthUserMixin, field='email') add_proc([encrypt_password], model=AuthUserMixin, field='password')
Set up event subscribers.
entailment
def _make_request(self, method, params): """Make request to API endpoint Note: Ignoring SSL cert validation due to intermittent failures http://requests.readthedocs.org/en/latest/user/advanced/#ssl-cert-verification """ if self.blocked_until is not None and \ datetime.datetime.utcnow() < self.blocked_until: raise SlackError("Too many requests - wait until {0}" \ .format(self.blocked_until)) url = "%s/%s" % (SlackClient.BASE_URL, method) params['token'] = self.token response = requests.post(url, data=params, verify=self.verify) if response.status_code == 429: # Too many requests retry_after = int(response.headers.get('retry-after', '1')) self.blocked_until = datetime.datetime.utcnow() + \ datetime.timedelta(seconds=retry_after) raise SlackError("Too many requests - retry after {0} second(s)" \ .format(retry_after)) result = response.json() if not result['ok']: raise SlackError(result['error']) return result
Make request to API endpoint Note: Ignoring SSL cert validation due to intermittent failures http://requests.readthedocs.org/en/latest/user/advanced/#ssl-cert-verification
entailment
def channels_list(self, exclude_archived=True, **params): """channels.list This method returns a list of all channels in the team. This includes channels the caller is in, channels they are not currently in, and archived channels. The number of (non-deactivated) members in each channel is also returned. https://api.slack.com/methods/channels.list """ method = 'channels.list' params.update({'exclude_archived': exclude_archived and 1 or 0}) return self._make_request(method, params)
channels.list This method returns a list of all channels in the team. This includes channels the caller is in, channels they are not currently in, and archived channels. The number of (non-deactivated) members in each channel is also returned. https://api.slack.com/methods/channels.list
entailment
def channel_name_to_id(self, channel_name, force_lookup=False): """Helper name for getting a channel's id from its name """ if force_lookup or not self.channel_name_id_map: channels = self.channels_list()['channels'] self.channel_name_id_map = {channel['name']: channel['id'] for channel in channels} channel = channel_name.startswith('#') and channel_name[1:] or channel_name return self.channel_name_id_map.get(channel)
Helper name for getting a channel's id from its name
entailment
def chat_post_message(self, channel, text, **params): """chat.postMessage This method posts a message to a channel. https://api.slack.com/methods/chat.postMessage """ method = 'chat.postMessage' params.update({ 'channel': channel, 'text': text, }) return self._make_request(method, params)
chat.postMessage This method posts a message to a channel. https://api.slack.com/methods/chat.postMessage
entailment
def chat_update_message(self, channel, text, timestamp, **params): """chat.update This method updates a message. Required parameters: `channel`: Channel containing the message to be updated. (e.g: "C1234567890") `text`: New text for the message, using the default formatting rules. (e.g: "Hello world") `timestamp`: Timestamp of the message to be updated (e.g: "1405894322.002768") https://api.slack.com/methods/chat.update """ method = 'chat.update' if self._channel_is_name(channel): # chat.update only takes channel ids (not channel names) channel = self.channel_name_to_id(channel) params.update({ 'channel': channel, 'text': text, 'ts': timestamp, }) return self._make_request(method, params)
chat.update This method updates a message. Required parameters: `channel`: Channel containing the message to be updated. (e.g: "C1234567890") `text`: New text for the message, using the default formatting rules. (e.g: "Hello world") `timestamp`: Timestamp of the message to be updated (e.g: "1405894322.002768") https://api.slack.com/methods/chat.update
entailment
def includeme(config): """ Connect view to route that catches all URIs like 'something,something,...' """ root = config.get_root_resource() root.add('nef_polymorphic', '{collections:.+,.+}', view=PolymorphicESView, factory=PolymorphicACL)
Connect view to route that catches all URIs like 'something,something,...'
entailment
def get_collections(self): """ Get names of collections from request matchdict. :return: Names of collections :rtype: list of str """ collections = self.request.matchdict['collections'].split('/')[0] collections = [coll.strip() for coll in collections.split(',')] return set(collections)
Get names of collections from request matchdict. :return: Names of collections :rtype: list of str
entailment
def get_resources(self, collections): """ Get resources that correspond to values from :collections:. :param collections: Collection names for which resources should be gathered :type collections: list of str :return: Gathered resources :rtype: list of Resource instances """ res_map = self.request.registry._model_collections resources = [res for res in res_map.values() if res.collection_name in collections] resources = [res for res in resources if res] return set(resources)
Get resources that correspond to values from :collections:. :param collections: Collection names for which resources should be gathered :type collections: list of str :return: Gathered resources :rtype: list of Resource instances
entailment
def _get_least_permissions_aces(self, resources): """ Get ACEs with the least permissions that fit all resources. To have access to polymorph on N collections, user MUST have access to all of them. If this is true, ACEs are returned, that allows 'view' permissions to current request principals. Otherwise None is returned thus blocking all permissions except those defined in `nefertari.acl.BaseACL`. :param resources: :type resources: list of Resource instances :return: Generated Pyramid ACEs or None :rtype: tuple or None """ factories = [res.view._factory for res in resources] contexts = [factory(self.request) for factory in factories] for ctx in contexts: if not self.request.has_permission('view', ctx): return else: return [ (Allow, principal, 'view') for principal in self.request.effective_principals ]
Get ACEs with the least permissions that fit all resources. To have access to polymorph on N collections, user MUST have access to all of them. If this is true, ACEs are returned, that allows 'view' permissions to current request principals. Otherwise None is returned thus blocking all permissions except those defined in `nefertari.acl.BaseACL`. :param resources: :type resources: list of Resource instances :return: Generated Pyramid ACEs or None :rtype: tuple or None
entailment
def set_collections_acl(self): """ Calculate and set ACL valid for requested collections. DENY_ALL is added to ACL to make sure no access rules are inherited. """ acl = [(Allow, 'g:admin', ALL_PERMISSIONS)] collections = self.get_collections() resources = self.get_resources(collections) aces = self._get_least_permissions_aces(resources) if aces is not None: for ace in aces: acl.append(ace) acl.append(DENY_ALL) self.__acl__ = tuple(acl)
Calculate and set ACL valid for requested collections. DENY_ALL is added to ACL to make sure no access rules are inherited.
entailment
def determine_types(self): """ Determine ES type names from request data. In particular `request.matchdict['collections']` is used to determine types names. Its value is comma-separated sequence of collection names under which views have been registered. """ from nefertari.elasticsearch import ES collections = self.get_collections() resources = self.get_resources(collections) models = set([res.view.Model for res in resources]) es_models = [mdl for mdl in models if mdl and getattr(mdl, '_index_enabled', False)] types = [ES.src2type(mdl.__name__) for mdl in es_models] return types
Determine ES type names from request data. In particular `request.matchdict['collections']` is used to determine types names. Its value is comma-separated sequence of collection names under which views have been registered.
entailment
def get_command(domain_name, command_name): """Returns a closure function that dispatches message to the WebSocket.""" def send_command(self, **kwargs): return self.ws.send_message( '{0}.{1}'.format(domain_name, command_name), kwargs ) return send_command
Returns a closure function that dispatches message to the WebSocket.
entailment
def DomainFactory(domain_name, cmds): """Dynamically create Domain class and set it's methods.""" klass = type(str(domain_name), (BaseDomain,), {}) for c in cmds: command = get_command(domain_name, c['name']) setattr(klass, c['name'], classmethod(command)) return klass
Dynamically create Domain class and set it's methods.
entailment
def main(): """The entrypoint for the hairball command installed via setup.py.""" description = ('PATH can be either the path to a scratch file, or a ' 'directory containing scratch files. Multiple PATH ' 'arguments can be provided.') parser = OptionParser(usage='%prog -p PLUGIN_NAME [options] PATH...', description=description, version='%prog {}'.format(__version__)) parser.add_option('-d', '--plugin-dir', metavar='DIR', help=('Specify the path to a directory containing ' 'plugins. Plugins in this directory take ' 'precedence over similarly named plugins ' 'included with Hairball.')) parser.add_option('-p', '--plugin', action='append', help=('Use the named plugin to perform analysis. ' 'This option can be provided multiple times.')) parser.add_option('-k', '--kurt-plugin', action='append', help=('Provide either a python import path (e.g, ' 'kelp.octopi) to a package/module, or the path' ' to a python file, which will be loaded as a ' 'Kurt plugin. This option can be provided ' 'multiple times.')) parser.add_option('-q', '--quiet', action='store_true', help=('Prevent output from Hairball. Plugins may still ' 'produce output.')) parser.add_option('-C', '--no-cache', action='store_true', help='Do not use Hairball\'s cache.', default=False) options, args = parser.parse_args(sys.argv[1:]) if not options.plugin: parser.error('At least one plugin must be specified via -p.') if not args: parser.error('At least one PATH must be provided.') if options.plugin_dir: if os.path.isdir(options.plugin_dir): sys.path.append(options.plugin_dir) else: parser.error('{} is not a directory'.format(options.plugin_dir)) hairball = Hairball(options, args, cache=not options.no_cache) hairball.initialize_plugins() hairball.process() hairball.finalize()
The entrypoint for the hairball command installed via setup.py.
entailment
def path_to_key(filepath): """Return the sha1sum (key) belonging to the file at filepath.""" tmp, last = os.path.split(filepath) tmp, middle = os.path.split(tmp) return '{}{}{}'.format(os.path.basename(tmp), middle, os.path.splitext(last)[0])
Return the sha1sum (key) belonging to the file at filepath.
entailment
def key_to_path(self, key): """Return the fullpath to the file with sha1sum key.""" return os.path.join(self.cache_dir, key[:2], key[2:4], key[4:] + '.pkl')
Return the fullpath to the file with sha1sum key.
entailment
def load(self, filename): """Optimized load and return the parsed version of filename. Uses the on-disk parse cache if the file is located in it. """ # Compute sha1 hash (key) with open(filename) as fp: key = sha1(fp.read()).hexdigest() path = self.key_to_path(key) # Return the cached file if available if key in self.hashes: try: with open(path) as fp: return cPickle.load(fp) except EOFError: os.unlink(path) self.hashes.remove(key) except IOError: self.hashes.remove(key) # Create the nested cache directory try: os.makedirs(os.path.dirname(path)) except OSError as exc: if exc.errno != errno.EEXIST: raise # Process the file and save in the cache scratch = kurt.Project.load(filename) # can fail with os.fdopen(os.open(path, os.O_WRONLY | os.O_CREAT, 0400), 'w') as fp: # open file for writing but make it immediately read-only cPickle.dump(scratch, fp, cPickle.HIGHEST_PROTOCOL) self.hashes.add(key) return scratch
Optimized load and return the parsed version of filename. Uses the on-disk parse cache if the file is located in it.
entailment
def hairball_files(self, paths, extensions): """Yield filepath to files with the proper extension within paths.""" def add_file(filename): return os.path.splitext(filename)[1] in extensions while paths: arg_path = paths.pop(0) if os.path.isdir(arg_path): found = False for path, dirs, files in os.walk(arg_path): dirs.sort() # Traverse in sorted order for filename in sorted(files): if add_file(filename): yield os.path.join(path, filename) found = True if not found: if not self.options.quiet: print('No files found in {}'.format(arg_path)) elif add_file(arg_path): yield arg_path elif not self.options.quiet: print('Invalid file {}'.format(arg_path)) print('Did you forget to load a Kurt plugin (-k)?')
Yield filepath to files with the proper extension within paths.
entailment
def initialize_plugins(self): """Attempt to Load and initialize all the plugins. Any issues loading plugins will be output to stderr. """ for plugin_name in self.options.plugin: parts = plugin_name.split('.') if len(parts) > 1: module_name = '.'.join(parts[:-1]) class_name = parts[-1] else: # Use the titlecase format of the module name as the class name module_name = parts[0] class_name = parts[0].title() # First try to load plugins from the passed in plugins_dir and then # from the hairball.plugins package. plugin = None for package in (None, 'hairball.plugins'): if package: module_name = '{}.{}'.format(package, module_name) try: module = __import__(module_name, fromlist=[class_name]) # Initializes the plugin by calling its constructor plugin = getattr(module, class_name)() # Verify plugin is of the correct class if not isinstance(plugin, HairballPlugin): sys.stderr.write('Invalid type for plugin {}: {}\n' .format(plugin_name, type(plugin))) plugin = None else: break except (ImportError, AttributeError): pass if plugin: self.plugins.append(plugin) else: sys.stderr.write('Cannot find plugin {}\n'.format(plugin_name)) if not self.plugins: sys.stderr.write('No plugins loaded. Goodbye!\n') sys.exit(1)
Attempt to Load and initialize all the plugins. Any issues loading plugins will be output to stderr.
entailment
def process(self): """Run the analysis across all files found in the given paths. Each file is loaded once and all plugins are run against it before loading the next file. """ for filename in self.hairball_files(self.paths, self.extensions): if not self.options.quiet: print(filename) try: if self.cache: scratch = self.cache.load(filename) else: scratch = kurt.Project.load(filename) except Exception: # pylint: disable=W0703 traceback.print_exc() continue for plugin in self.plugins: # pylint: disable=W0212 plugin._process(scratch, filename=filename)
Run the analysis across all files found in the given paths. Each file is loaded once and all plugins are run against it before loading the next file.
entailment
def maybe_dotted(module, throw=True): """ If ``module`` is a dotted string pointing to the module, imports and returns the module object. """ try: return Configurator().maybe_dotted(module) except ImportError as e: err = '%s not found. %s' % (module, e) if throw: raise ImportError(err) else: log.error(err) return None
If ``module`` is a dotted string pointing to the module, imports and returns the module object.
entailment
def issequence(arg): """Return True if `arg` acts as a list and does not look like a string.""" string_behaviour = ( isinstance(arg, six.string_types) or isinstance(arg, six.text_type)) list_behaviour = hasattr(arg, '__getitem__') or hasattr(arg, '__iter__') return not string_behaviour and list_behaviour
Return True if `arg` acts as a list and does not look like a string.
entailment
def merge_dicts(a, b, path=None): """ Merge dict :b: into dict :a: Code snippet from http://stackoverflow.com/a/7205107 """ if path is None: path = [] for key in b: if key in a: if isinstance(a[key], dict) and isinstance(b[key], dict): merge_dicts(a[key], b[key], path + [str(key)]) elif a[key] == b[key]: pass # same leaf value else: raise Exception( 'Conflict at %s' % '.'.join(path + [str(key)])) else: a[key] = b[key] return a
Merge dict :b: into dict :a: Code snippet from http://stackoverflow.com/a/7205107
entailment
def str2dict(dotted_str, value=None, separator='.'): """ Convert dotted string to dict splitting by :separator: """ dict_ = {} parts = dotted_str.split(separator) d, prev = dict_, None for part in parts: prev = d d = d.setdefault(part, {}) else: if value is not None: prev[part] = value return dict_
Convert dotted string to dict splitting by :separator:
entailment
def validate_data_privacy(request, data, wrapper_kw=None): """ Validate :data: contains only data allowed by privacy settings. :param request: Pyramid Request instance :param data: Dict containing request/response data which should be validated """ from nefertari import wrappers if wrapper_kw is None: wrapper_kw = {} wrapper = wrappers.apply_privacy(request) allowed_fields = wrapper(result=data, **wrapper_kw).keys() data = data.copy() data.pop('_type', None) not_allowed_fields = set(data.keys()) - set(allowed_fields) if not_allowed_fields: raise wrappers.ValidationError(', '.join(not_allowed_fields))
Validate :data: contains only data allowed by privacy settings. :param request: Pyramid Request instance :param data: Dict containing request/response data which should be validated
entailment
def drop_reserved_params(params): """ Drops reserved params """ from nefertari import RESERVED_PARAMS params = params.copy() for reserved_param in RESERVED_PARAMS: if reserved_param in params: params.pop(reserved_param) return params
Drops reserved params
entailment
def _send_command_raw(self, command, opt=''): """ Description: The TV doesn't handle long running connections very well, so we open a new connection every time. There might be a better way to do this, but it's pretty quick and resilient. Returns: If a value is being requested ( opt2 is "?" ), then the return value is returned. If a value is being set, it returns True for "OK" or False for "ERR" """ # According to the documentation: # http://files.sharpusa.com/Downloads/ForHome/ # HomeEntertainment/LCDTVs/Manuals/tel_man_LC40_46_52_60LE830U.pdf # Page 58 - Communication conditions for IP # The connection could be lost (but not only after 3 minutes), # so we need to the remote commands to be sure about states end_time = time.time() + self.timeout while time.time() < end_time: try: # Connect sock_con = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock_con.settimeout(self.connection_timeout) sock_con.connect((self.ip_address, self.port)) # Authenticate sock_con.send(self.auth) sock_con.recv(1024) sock_con.recv(1024) # Send command if opt != '': command += str(opt) sock_con.send(str.encode(command.ljust(8) + '\r')) status = bytes.decode(sock_con.recv(1024)).strip() except (OSError, socket.error) as exp: time.sleep(0.1) if time.time() >= end_time: raise exp else: sock_con.close() # Sometimes the status is empty so # We need to retry if status != u'': break if status == "OK": return True elif status == "ERR": return False else: try: return int(status) except ValueError: return status
Description: The TV doesn't handle long running connections very well, so we open a new connection every time. There might be a better way to do this, but it's pretty quick and resilient. Returns: If a value is being requested ( opt2 is "?" ), then the return value is returned. If a value is being set, it returns True for "OK" or False for "ERR"
entailment
def info(self): """ Description: Returns dict of information about the TV name, model, version """ return {"name": self._send_command('name'), "model": self._send_command('model'), "version": self._send_command('version'), "ip_version": self._send_command('ip_version') }
Description: Returns dict of information about the TV name, model, version
entailment
def get_input_list(self): """ Description: Get input list Returns an ordered list of all available input keys and names """ inputs = [' '] * len(self.command['input']) for key in self.command['input']: inputs[self.command['input'][key]['order']] = {"key":key, "name":self.command['input'][key]['name']} return inputs
Description: Get input list Returns an ordered list of all available input keys and names
entailment
def input(self, opt): """ Description: Set the input Call with no arguments to get current setting Arguments: opt: string Name provided from input list or key from yaml ("HDMI 1" or "hdmi_1") """ for key in self.command['input']: if (key == opt) or (self.command['input'][key]['name'] == opt): return self._send_command(['input', key, 'command']) return False
Description: Set the input Call with no arguments to get current setting Arguments: opt: string Name provided from input list or key from yaml ("HDMI 1" or "hdmi_1")
entailment
def digital_channel_air(self, opt1='?', opt2='?'): """ Description: Change Channel (Digital) Pass Channels "XX.YY" as TV.digital_channel_air(XX, YY) Arguments: opt1: integer 1-99: Major Channel opt2: integer (optional) 1-99: Minor Channel """ if opt1 == '?': parameter = '?' elif opt2 == '?': parameter = str(opt1).rjust(4, "0") else: parameter = '{:02d}{:02d}'.format(opt1, opt2) return self._send_command('digital_channel_air', parameter)
Description: Change Channel (Digital) Pass Channels "XX.YY" as TV.digital_channel_air(XX, YY) Arguments: opt1: integer 1-99: Major Channel opt2: integer (optional) 1-99: Minor Channel
entailment
def digital_channel_cable(self, opt1='?', opt2=0): """ Description: Change Channel (Digital) Pass Channels "XXX.YYY" as TV.digital_channel_cable(XXX, YYY) Arguments: opt1: integer 1-999: Major Channel opt2: integer (optional) 0-999: Minor Channel """ if opt1 == '?': parameter = '?' elif self.command['digital_channel_cable_minor'] == '': parameter = str(opt1).rjust(4, "0") else: self._send_command('digital_channel_cable_minor', str(opt1).rjust(3, "0")) parameter = str(opt2).rjust(3, "0") return self._send_command('digital_channel_cable_major', parameter)
Description: Change Channel (Digital) Pass Channels "XXX.YYY" as TV.digital_channel_cable(XXX, YYY) Arguments: opt1: integer 1-999: Major Channel opt2: integer (optional) 0-999: Minor Channel
entailment
def get_remote_button_list(self): """ Description: Get remote button list Returns an list of all available remote buttons """ remote_buttons = [] for key in self.command['remote']: if self.command['remote'][key] != '': remote_buttons.append(key) return remote_buttons
Description: Get remote button list Returns an list of all available remote buttons
entailment
def check_results(tmp_): """Return a 3 tuple for something.""" # TODO: Fix this to work with more meaningful names if tmp_['t'] > 0: if tmp_['l'] > 0: if tmp_['rr'] > 0 or tmp_['ra'] > 1: print 1, 3, tmp_ return 3 elif tmp_['cr'] > 0 or tmp_['ca'] > 1: print 2, 3, tmp_ return 3 elif tmp_['mr'] > 0 or tmp_['ma'] > 1: print 3, 2, tmp_ return 2 if tmp_['cr'] > 1 or tmp_['ca'] > 2: print 4, 2, tmp_ return 2 if tmp_['mr'] > 0 or tmp_['ma'] > 1: if tmp_['cr'] > 0 or tmp_['ca'] > 1: print 6, 0, tmp_ return 0 if tmp_['rr'] > 1 or tmp_['ra'] > 2: print 7, 0, tmp_ return 0 if tmp_['sr'] > 1 or tmp_['sa'] > 2: print 8, 0, tmp_ return 0 if tmp_['l'] > 0: if tmp_['rr'] > 0 or tmp_['ra'] > 1: print 9, 2, tmp_ return 2 if tmp_['cr'] > 0 or tmp_['ca'] > 1: print 10, 0, tmp_ return 0 return -1
Return a 3 tuple for something.
entailment
def _check_animation(self, last, last_level, gen): """Internal helper function to check the animation.""" tmp_ = Counter() results = Counter() name, level, block = last, last_level, last others = False while name in self.ANIMATION and level >= last_level: if name in self.LOOP: if block != last: count = self.check_results(tmp_) if count > -1: results[count] += 1 tmp_.clear() tmp_['last'] += 1 for attribute in ('costume', 'orientation', 'position', 'size'): if (name, 'relative') in self.BLOCKMAPPING[attribute]: tmp_[(attribute, 'relative')] += 1 elif (name, 'absolute') in self.BLOCKMAPPING[attribute]: tmp_[(attribute, 'absolute')] += 1 if name in self.TIMING: tmp_['timing'] += 1 last_level = level name, level, block = next(gen, ('', 0, '')) # allow some exceptions if name not in self.ANIMATION and name != '': if not others: if block.type.shape != 'stack': last_level = level (name, level, block) = next(gen, ('', 0, '')) others = True count = self.check_results(tmp_) if count > -1: results[count] += 1 return gen, results
Internal helper function to check the animation.
entailment
def analyze(self, scratch, **kwargs): """Run and return the results from the Animation plugin.""" results = Counter() for script in self.iter_scripts(scratch): gen = self.iter_blocks(script.blocks) name = 'start' level = None while name != '': if name in self.ANIMATION: gen, count = self._check_animation(name, level, gen) results.update(count) name, level, _ = next(gen, ('', 0, '')) return {'animation': results}
Run and return the results from the Animation plugin.
entailment
def get_receive(self, script_list): """Return a list of received events contained in script_list.""" events = defaultdict(set) for script in script_list: if self.script_start_type(script) == self.HAT_WHEN_I_RECEIVE: event = script.blocks[0].args[0].lower() events[event].add(script) return events
Return a list of received events contained in script_list.
entailment
def analyze(self, scratch, **kwargs): """Run and return the results from the BroadcastReceive plugin.""" all_scripts = list(self.iter_scripts(scratch)) results = defaultdict(set) broadcast = dict((x, self.get_broadcast_events(x)) # Events by script for x in all_scripts) correct = self.get_receive(all_scripts) results['never broadcast'] = set(correct.keys()) for script, events in broadcast.items(): for event in events.keys(): if event is True: # Remove dynamic broadcasts results['dynamic broadcast'].add(script.morph.name) del events[event] elif event in correct: results['never broadcast'].discard(event) else: results['never received'].add(event) # remove events from correct dict that were never broadcast for event in correct.keys(): if event in results['never broadcast']: del correct[event] # Find scripts that have more than one broadcast event on any possible # execution path through the program # TODO: Permit mutually exclusive broadcasts for events in broadcast.values(): if len(events) > 1: for event in events: if event in correct: results['parallel broadcasts'].add(event) del correct[event] # Find events that have two (or more) receivers in which one of the # receivers has a "delay" block for event, scripts in correct.items(): if len(scripts) > 1: for script in scripts: for _, _, block in self.iter_blocks(script.blocks): if block.type.shape == 'stack': results['multiple receivers with delay'].add(event) if event in correct: del correct[event] results['success'] = set(correct.keys()) return {'broadcast': results}
Run and return the results from the BroadcastReceive plugin.
entailment
def analyze(self, scratch, **kwargs): """Categorize instances of attempted say and sound synchronization.""" errors = Counter() for script in self.iter_scripts(scratch): prev_name, prev_depth, prev_block = '', 0, script.blocks[0] gen = self.iter_blocks(script.blocks) for name, depth, block in gen: if prev_depth == depth: if prev_name in self.SAY_THINK: if name == 'play sound %s until done': if not self.is_blank(prev_block.args[0]): errors += self.check(gen) # TODO: What about play sound? elif prev_name in self.SAY_THINK_DURATION and \ 'play sound %s' in name: errors['1'] += 1 elif prev_name == 'play sound %s': if name in self.SAY_THINK: errors[self.INCORRECT] += 1 elif name in self.SAY_THINK_DURATION: if self.is_blank(block.args[0]): errors[self.ERROR] += 1 else: errors[self.HACKISH] += 1 elif prev_name == 'play sound %s until done' and \ name in self.ALL_SAY_THINK: if not self.is_blank(block.args[0]): errors[self.INCORRECT] += 1 # TODO: Should there be an else clause here? prev_name, prev_depth, prev_block = name, depth, block return {'sound': errors}
Categorize instances of attempted say and sound synchronization.
entailment
def check(self, gen): """Check that the last part of the chain matches. TODO: Fix to handle the following situation that appears to not work say 'message 1' play sound until done say 'message 2' say 'message 3' play sound until done say '' """ retval = Counter() name, _, block = next(gen, ('', 0, '')) if name in self.SAY_THINK: if self.is_blank(block.args[0]): retval[self.CORRECT] += 1 else: name, _, block = next(gen, ('', 0, '')) if name == 'play sound %s until done': # Increment the correct count because we have at least # one successful instance retval[self.CORRECT] += 1 # This block represents the beginning of a second retval += self.check(gen) else: retval[self.INCORRECT] += 1 else: retval[self.INCORRECT] += 1 return retval
Check that the last part of the chain matches. TODO: Fix to handle the following situation that appears to not work say 'message 1' play sound until done say 'message 2' say 'message 3' play sound until done say ''
entailment
def _format_exception(err, is_failure, stdout=None, stderr=None): """Converts a sys.exc_info()-style tuple of values into a string.""" exctype, value, tb = err # Skip test runner traceback levels while tb and _is_relevant_tb_level(tb): tb = tb.tb_next if is_failure: # Skip assert*() traceback levels length = _count_relevant_tb_levels(tb) msgLines = traceback.format_exception(exctype, value, tb, length) else: msgLines = traceback.format_exception(exctype, value, tb) encoding = locale.getpreferredencoding() msgLines = [_decode(line, encoding) for line in msgLines] if stdout: if not stdout.endswith('\n'): stdout += '\n' msgLines.append(STDOUT_LINE % stdout) if stderr: if not stderr.endswith('\n'): stderr += '\n' msgLines.append(STDERR_LINE % stderr) return ''.join(msgLines)
Converts a sys.exc_info()-style tuple of values into a string.
entailment
def _setup_stdout(self): """Hook stdout and stderr if buffering is enabled. """ if self.buffer: if self._stderr_buffer is None: self._stderr_buffer = StringIO() self._stdout_buffer = StringIO() sys.stdout = self._stdout_buffer sys.stderr = self._stderr_buffer
Hook stdout and stderr if buffering is enabled.
entailment
def _restore_stdout(self): """Unhook stdout and stderr if buffering is enabled. """ if self.buffer: if self._mirror_output: output = sys.stdout.getvalue() error = sys.stderr.getvalue() if output: if not output.endswith('\n'): output += '\n' self._original_stdout.write(STDOUT_LINE % output) if error: if not error.endswith('\n'): error += '\n' self._original_stderr.write(STDERR_LINE % error) sys.stdout = self._original_stdout sys.stderr = self._original_stderr self._stdout_buffer.seek(0) self._stdout_buffer.truncate() self._stderr_buffer.seek(0) self._stderr_buffer.truncate()
Unhook stdout and stderr if buffering is enabled.
entailment
def add_result_handler(self, handler): """Register a new result handler. """ self._result_handlers.append(handler) # Reset sorted handlers if self._sorted_handlers: self._sorted_handlers = None
Register a new result handler.
entailment
def add_result(self, result): """Add an already-constructed :class:`~.TestResult` to this :class:`~.ResultCollector`. This may be used when collecting results created by other ResultCollectors (e.g. in subprocesses). """ for handler in self._handlers: handler(result) if self._successful and result.status not in _successful_results: self._successful = False
Add an already-constructed :class:`~.TestResult` to this :class:`~.ResultCollector`. This may be used when collecting results created by other ResultCollectors (e.g. in subprocesses).
entailment
def _handle_result(self, test, status, exception=None, message=None): """Create a :class:`~.TestResult` and add it to this :class:`~ResultCollector`. Parameters ---------- test : unittest.TestCase The test that this result will represent. status : haas.result.TestCompletionStatus The status of the test. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``. message : str Optional message associated with the result (e.g. skip reason). """ if self.buffer: stderr = self._stderr_buffer.getvalue() stdout = self._stdout_buffer.getvalue() else: stderr = stdout = None started_time = self._test_timing.get(self._testcase_to_key(test)) if started_time is None and isinstance(test, ErrorHolder): started_time = datetime.utcnow() elif started_time is None: raise RuntimeError( 'Missing test start! Please report this error as a bug in ' 'haas.') completion_time = datetime.utcnow() duration = TestDuration(started_time, completion_time) result = TestResult.from_test_case( test, status, duration=duration, exception=exception, message=message, stdout=stdout, stderr=stderr, ) self.add_result(result) return result
Create a :class:`~.TestResult` and add it to this :class:`~ResultCollector`. Parameters ---------- test : unittest.TestCase The test that this result will represent. status : haas.result.TestCompletionStatus The status of the test. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``. message : str Optional message associated with the result (e.g. skip reason).
entailment
def addError(self, test, exception): """Register that a test ended in an error. Parameters ---------- test : unittest.TestCase The test that has completed. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``. """ result = self._handle_result( test, TestCompletionStatus.error, exception=exception) self.errors.append(result) self._mirror_output = True
Register that a test ended in an error. Parameters ---------- test : unittest.TestCase The test that has completed. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``.
entailment
def addFailure(self, test, exception): """Register that a test ended with a failure. Parameters ---------- test : unittest.TestCase The test that has completed. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``. """ result = self._handle_result( test, TestCompletionStatus.failure, exception=exception) self.failures.append(result) self._mirror_output = True
Register that a test ended with a failure. Parameters ---------- test : unittest.TestCase The test that has completed. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``.
entailment
def addSkip(self, test, reason): """Register that a test that was skipped. Parameters ---------- test : unittest.TestCase The test that has completed. reason : str The reason the test was skipped. """ result = self._handle_result( test, TestCompletionStatus.skipped, message=reason) self.skipped.append(result)
Register that a test that was skipped. Parameters ---------- test : unittest.TestCase The test that has completed. reason : str The reason the test was skipped.
entailment
def addExpectedFailure(self, test, exception): """Register that a test that failed and was expected to fail. Parameters ---------- test : unittest.TestCase The test that has completed. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``. """ result = self._handle_result( test, TestCompletionStatus.expected_failure, exception=exception) self.expectedFailures.append(result)
Register that a test that failed and was expected to fail. Parameters ---------- test : unittest.TestCase The test that has completed. exception : tuple ``exc_info`` tuple ``(type, value, traceback)``.
entailment
def addUnexpectedSuccess(self, test): """Register a test that passed unexpectedly. Parameters ---------- test : unittest.TestCase The test that has completed. """ result = self._handle_result( test, TestCompletionStatus.unexpected_success) self.unexpectedSuccesses.append(result)
Register a test that passed unexpectedly. Parameters ---------- test : unittest.TestCase The test that has completed.
entailment