INSTRUCTION
stringlengths
1
8.43k
RESPONSE
stringlengths
75
104k
Check arguments passed by user that are not checked by argparse itself.
def check_arguments(args, parser): """Check arguments passed by user that are not checked by argparse itself.""" if args.asm_block not in ['auto', 'manual']: try: args.asm_block = int(args.asm_block) except ValueError: parser.error('--asm-block can only be "auto", "manual" or an integer') # Set default unit depending on performance model requested if not args.unit: if 'Roofline' in args.pmodel or 'RooflineIACA' in args.pmodel: args.unit = 'FLOP/s' else: args.unit = 'cy/CL'
Run command line interface.
def run(parser, args, output_file=sys.stdout): """Run command line interface.""" # Try loading results file (if requested) result_storage = {} if args.store: args.store.seek(0) try: result_storage = pickle.load(args.store) except EOFError: pass args.store.close() # machine information # Read machine description machine = MachineModel(args.machine.name, args=args) # process kernel if not args.kernel_description: code = str(args.code_file.read()) code = clean_code(code) kernel = KernelCode(code, filename=args.code_file.name, machine=machine, keep_intermediates=not args.clean_intermediates) else: description = str(args.code_file.read()) kernel = KernelDescription(yaml.load(description, Loader=yaml.Loader), machine=machine) # if no defines were given, guess suitable defines in-mem # TODO support in-cache # TODO broaden cases to n-dimensions # TODO make configurable (no hardcoded 512MB/1GB/min. 3 iteration ...) # works only for up to 3 dimensions required_consts = [v[1] for v in kernel.variables.values() if v[1] is not None] required_consts += [[l['start'], l['stop']] for l in kernel.get_loop_stack()] # split into individual consts required_consts = [i for l in required_consts for i in l] required_consts = set([i for l in required_consts for i in l.free_symbols]) if len(required_consts) > 0: # build defines permutations define_dict = {} for name, values in args.define: if name not in define_dict: define_dict[name] = [[name, v] for v in values] continue for v in values: if v not in define_dict[name]: define_dict[name].append([name, v]) define_product = list(itertools.product(*list(define_dict.values()))) # Check that all consts have been defined if set(required_consts).difference(set([symbol_pos_int(k) for k in define_dict.keys()])): raise ValueError("Not all constants have been defined. Required are: {}".format( required_consts)) else: define_product = [{}] for define in define_product: # Reset state of kernel kernel.clear_state() # Add constants from define arguments for k, v in define: kernel.set_constant(k, v) for model_name in uniquify(args.pmodel): # print header print('{:^80}'.format(' kerncraft '), file=output_file) print('{:<40}{:>40}'.format(args.code_file.name, '-m ' + args.machine.name), file=output_file) print(' '.join(['-D {} {}'.format(k, v) for k, v in define]), file=output_file) print('{:-^80}'.format(' ' + model_name + ' '), file=output_file) if args.verbose > 1: if not args.kernel_description: kernel.print_kernel_code(output_file=output_file) print('', file=output_file) kernel.print_variables_info(output_file=output_file) kernel.print_kernel_info(output_file=output_file) if args.verbose > 0: kernel.print_constants_info(output_file=output_file) model = getattr(models, model_name)(kernel, machine, args, parser) model.analyze() model.report(output_file=output_file) # Add results to storage kernel_name = os.path.split(args.code_file.name)[1] if kernel_name not in result_storage: result_storage[kernel_name] = {} if tuple(kernel.constants.items()) not in result_storage[kernel_name]: result_storage[kernel_name][tuple(kernel.constants.items())] = {} result_storage[kernel_name][tuple(kernel.constants.items())][model_name] = \ model.results print('', file=output_file) # Save storage to file (if requested) if args.store: temp_name = args.store.name + '.tmp' with open(temp_name, 'wb+') as f: pickle.dump(result_storage, f) shutil.move(temp_name, args.store.name)
Initialize and run command line interface.
def main(): """Initialize and run command line interface.""" # Create and populate parser parser = create_parser() # Parse given arguments args = parser.parse_args() # Checking arguments check_arguments(args, parser) # BUSINESS LOGIC IS FOLLOWING run(parser, args)
Comand line interface of picklemerge.
def main(): """Comand line interface of picklemerge.""" parser = argparse.ArgumentParser( description='Recursively merges two or more pickle files. Only supports pickles consisting ' 'of a single dictionary object.') parser.add_argument('destination', type=argparse.FileType('r+b'), help='File to write to and include in resulting pickle. (WILL BE CHANGED)') parser.add_argument('source', type=argparse.FileType('rb'), nargs='+', help='File to include in resulting pickle.') args = parser.parse_args() result = pickle.load(args.destination) assert isinstance(result, collections.Mapping), "only Mapping types can be handled." for s in args.source: data = pickle.load(s) assert isinstance(data, collections.Mapping), "only Mapping types can be handled." update(result, data) args.destination.seek(0) args.destination.truncate() pickle.dump(result, args.destination)
Create a sympy. Symbol with positive and integer assumptions.
def symbol_pos_int(*args, **kwargs): """Create a sympy.Symbol with positive and integer assumptions.""" kwargs.update({'positive': True, 'integer': True}) return sympy.Symbol(*args, **kwargs)
Prefix and indent all lines in * textblock *.
def prefix_indent(prefix, textblock, later_prefix=' '): """ Prefix and indent all lines in *textblock*. *prefix* is a prefix string *later_prefix* is used on all but the first line, if it is a single character it will be repeated to match length of *prefix* """ textblock = textblock.split('\n') line = prefix + textblock[0] + '\n' if len(later_prefix) == 1: later_prefix = ' '*len(prefix) line = line + '\n'.join([later_prefix + x for x in textblock[1:]]) if line[-1] != '\n': return line + '\n' else: return line
Transform ast of multidimensional declaration to a single dimension declaration.
def transform_multidim_to_1d_decl(decl): """ Transform ast of multidimensional declaration to a single dimension declaration. In-place operation! Returns name and dimensions of array (to be used with transform_multidim_to_1d_ref()) """ dims = [] type_ = decl.type while type(type_) is c_ast.ArrayDecl: dims.append(type_.dim) type_ = type_.type if dims: # Multidimensional array decl.type.dim = reduce(lambda l, r: c_ast.BinaryOp('*', l, r), dims) decl.type.type = type_ return decl.name, dims
Transform ast of multidimensional reference to a single dimension reference.
def transform_multidim_to_1d_ref(aref, dimension_dict): """ Transform ast of multidimensional reference to a single dimension reference. In-place operation! """ dims = [] name = aref while type(name) is c_ast.ArrayRef: dims.append(name.subscript) name = name.name subscript_list = [] for i, d in enumerate(dims): if i == 0: subscript_list.append(d) else: subscript_list.append(c_ast.BinaryOp('*', d, reduce( lambda l, r: c_ast.BinaryOp('*', l, r), dimension_dict[name.name][-1:-i-1:-1]))) aref.subscript = reduce( lambda l, r: c_ast.BinaryOp('+', l, r), subscript_list) aref.name = name
Transform ast of type var_name [ N ] to type * var_name = aligned_malloc ( sizeof ( type ) * N 32 )
def transform_array_decl_to_malloc(decl, with_init=True): """ Transform ast of "type var_name[N]" to "type* var_name = aligned_malloc(sizeof(type)*N, 32)" In-place operation. :param with_init: if False, ommit malloc """ if type(decl.type) is not c_ast.ArrayDecl: # Not an array declaration, can be ignored return type_ = c_ast.PtrDecl([], decl.type.type) if with_init: decl.init = c_ast.FuncCall( c_ast.ID('aligned_malloc'), c_ast.ExprList([ c_ast.BinaryOp( '*', c_ast.UnaryOp( 'sizeof', c_ast.Typename(None, [], c_ast.TypeDecl( None, [], decl.type.type.type))), decl.type.dim), c_ast.Constant('int', '32')])) decl.type = type_
Return list of array references in AST.
def find_node_type(ast, node_type): """Return list of array references in AST.""" if type(ast) is node_type: return [ast] elif type(ast) is list: return reduce(operator.add, list(map(lambda a: find_node_type(a, node_type), ast)), []) elif ast is None: return [] else: return reduce(operator.add, [find_node_type(o[1], node_type) for o in ast.children()], [])
Will make any functions return an iterable objects by wrapping its result in a list.
def force_iterable(f): """Will make any functions return an iterable objects by wrapping its result in a list.""" def wrapper(*args, **kwargs): r = f(*args, **kwargs) if hasattr(r, '__iter__'): return r else: return [r] return wrapper
Reduce absolute path to relative ( if shorter ) for easier readability.
def reduce_path(path): """Reduce absolute path to relative (if shorter) for easier readability.""" relative_path = os.path.relpath(path) if len(relative_path) < len(path): return relative_path else: return path
Check that information about kernel makes sens and is valid.
def check(self): """Check that information about kernel makes sens and is valid.""" datatypes = [v[0] for v in self.variables.values()] assert len(set(datatypes)) <= 1, 'mixing of datatypes within a kernel is not supported.'
Set constant of name to value.
def set_constant(self, name, value): """ Set constant of name to value. :param name: may be a str or a sympy.Symbol :param value: must be an int """ assert isinstance(name, str) or isinstance(name, sympy.Symbol), \ "constant name needs to be of type str, unicode or a sympy.Symbol" assert type(value) is int, "constant value needs to be of type int" if isinstance(name, sympy.Symbol): self.constants[name] = value else: self.constants[symbol_pos_int(name)] = value
Register variable of name and type_ with a ( multidimensional ) size.
def set_variable(self, name, type_, size): """ Register variable of name and type_, with a (multidimensional) size. :param name: variable name as it appears in code :param type_: may be any key from Kernel.datatypes_size (typically float or double) :param size: either None for scalars or an n-tuple of ints for an n-dimensional array """ assert type_ in self.datatypes_size, 'only float and double variables are supported' if self.datatype is None: self.datatype = type_ else: assert type_ == self.datatype, 'mixing of datatypes within a kernel is not supported.' assert type(size) in [tuple, type(None)], 'size has to be defined as tuple or None' self.variables[name] = (type_, size)
Substitute constants in expression unless it is already a number.
def subs_consts(self, expr): """Substitute constants in expression unless it is already a number.""" if isinstance(expr, numbers.Number): return expr else: return expr.subs(self.constants)
Return a dictionary with all arrays sizes.
def array_sizes(self, in_bytes=False, subs_consts=False): """ Return a dictionary with all arrays sizes. :param in_bytes: If True, output will be in bytes, not element counts. :param subs_consts: If True, output will be numbers and not symbolic. Scalar variables are ignored. """ var_sizes = {} for var_name, var_info in self.variables.items(): var_type, var_size = var_info # Skiping sclars if var_size is None: continue var_sizes[var_name] = reduce(operator.mul, var_size, 1) # Multiply by bytes per element if requested if in_bytes: element_size = self.datatypes_size[var_type] var_sizes[var_name] *= element_size if subs_consts: return {k: self.subs_consts(v) for k, v in var_sizes.items()} else: return var_sizes
Return the offset from the iteration center in number of elements.
def _calculate_relative_offset(self, name, access_dimensions): """ Return the offset from the iteration center in number of elements. The order of indices used in access is preserved. """ # TODO to be replaced with compile_global_offsets offset = 0 base_dims = self.variables[name][1] for dim, offset_info in enumerate(access_dimensions): offset_type, idx_name, dim_offset = offset_info assert offset_type == 'rel', 'Only relative access to arrays is supported at the moment' if offset_type == 'rel': offset += self.subs_consts( dim_offset*reduce(operator.mul, base_dims[dim+1:], sympy.Integer(1))) else: # should not happen pass return offset
Remove duplicate source and destination accesses
def _remove_duplicate_accesses(self): """ Remove duplicate source and destination accesses """ self.destinations = {var_name: set(acs) for var_name, acs in self.destinations.items()} self.sources = {var_name: set(acs) for var_name, acs in self.sources.items()}
Transform a ( multidimensional ) variable access to a flattend sympy expression.
def access_to_sympy(self, var_name, access): """ Transform a (multidimensional) variable access to a flattend sympy expression. Also works with flat array accesses. """ base_sizes = self.variables[var_name][1] expr = sympy.Number(0) for dimension, a in enumerate(access): base_size = reduce(operator.mul, base_sizes[dimension+1:], sympy.Integer(1)) expr += base_size*a return expr
Return the number of global loop iterations that are performed.
def iteration_length(self, dimension=None): """ Return the number of global loop iterations that are performed. If dimension is not None, it is the loop dimension that is returned (-1 is the inner most loop and 0 the outermost) """ total_length = 1 if dimension is not None: loops = [self._loop_stack[dimension]] else: loops = reversed(self._loop_stack) for var_name, start, end, incr in loops: # This unspools the iterations: length = end-start total_length = total_length*length return self.subs_consts(total_length)
Yield loop stack dictionaries in order from outer to inner.
def get_loop_stack(self, subs_consts=False): """Yield loop stack dictionaries in order from outer to inner.""" for l in self._loop_stack: if subs_consts: yield {'index': l[0], 'start': self.subs_consts(l[1]), 'stop': self.subs_consts(l[2]), 'increment': self.subs_consts(l[3])} else: yield {'index': l[0], 'start': l[1], 'stop': l[2], 'increment': l[3]}
Return the order of indices as they appear in array references.
def index_order(self, sources=True, destinations=True): """ Return the order of indices as they appear in array references. Use *source* and *destination* to filter output """ if sources: arefs = chain(*self.sources.values()) else: arefs = [] if destinations: arefs = chain(arefs, *self.destinations.values()) ret = [] for a in [aref for aref in arefs if aref is not None]: ref = [] for expr in a: ref.append(expr.free_symbols) ret.append(ref) return ret
Return a dictionary of lists of sympy accesses for each variable.
def compile_sympy_accesses(self, sources=True, destinations=True): """ Return a dictionary of lists of sympy accesses, for each variable. Use *source* and *destination* to filter output """ sympy_accesses = defaultdict(list) # Compile sympy accesses for var_name in self.variables: if sources: for r in self.sources.get(var_name, []): if r is None: continue sympy_accesses[var_name].append(self.access_to_sympy(var_name, r)) if destinations: for w in self.destinations.get(var_name, []): if w is None: continue sympy_accesses[var_name].append(self.access_to_sympy(var_name, w)) return sympy_accesses
Return load and store distances between accesses.
def compile_relative_distances(self, sympy_accesses=None): """ Return load and store distances between accesses. :param sympy_accesses: optionally restrict accesses, default from compile_sympy_accesses() e.g. if accesses are to [+N, +1, -1, -N], relative distances are [N-1, 2, N-1] returned is a dict of list of sympy expressions, for each variable """ if sympy_accesses is None: sympy_accesses = self.compile_sympy_accesses() sympy_distances = defaultdict(list) for var_name, accesses in sympy_accesses.items(): for i in range(1, len(accesses)): sympy_distances[var_name].append((accesses[i-1]-accesses[i]).simplify()) return sympy_distances
Return sympy expressions translating global_iterator to loop indices.
def global_iterator_to_indices(self, git=None): """ Return sympy expressions translating global_iterator to loop indices. If global_iterator is given, an integer is returned """ # unwind global iteration count into loop counters: base_loop_counters = {} global_iterator = symbol_pos_int('global_iterator') idiv = implemented_function(sympy.Function(str('idiv')), lambda x, y: x//y) total_length = 1 last_incr = 1 for var_name, start, end, incr in reversed(self._loop_stack): loop_var = symbol_pos_int(var_name) # This unspools the iterations: length = end-start # FIXME is incr handled correct here? counter = start+(idiv(global_iterator*last_incr, total_length)*incr) % length total_length = total_length*length last_incr = incr base_loop_counters[loop_var] = sympy.lambdify( global_iterator, self.subs_consts(counter), modules=[numpy, {'Mod': numpy.mod}]) if git is not None: try: # Try to resolve to integer if global_iterator was given base_loop_counters[loop_var] = sympy.Integer(self.subs_consts(counter)) continue except (ValueError, TypeError): base_loop_counters[loop_var] = base_loop_counters[loop_var](git) return base_loop_counters
Return global iterator sympy expression
def global_iterator(self): """ Return global iterator sympy expression """ global_iterator = sympy.Integer(0) total_length = sympy.Integer(1) for var_name, start, end, incr in reversed(self._loop_stack): loop_var = symbol_pos_int(var_name) length = end - start # FIXME is incr handled correct here? global_iterator += (loop_var - start) * total_length total_length *= length return global_iterator
Transform a dictionary of indices to a global iterator integer.
def indices_to_global_iterator(self, indices): """ Transform a dictionary of indices to a global iterator integer. Inverse of global_iterator_to_indices(). """ global_iterator = self.subs_consts(self.global_iterator().subs(indices)) return global_iterator
Return global iterator with last iteration number
def max_global_iteration(self): """Return global iterator with last iteration number""" return self.indices_to_global_iterator({ symbol_pos_int(var_name): end-1 for var_name, start, end, incr in self._loop_stack })
Return load and store offsets on a virtual address space.
def compile_global_offsets(self, iteration=0, spacing=0): """ Return load and store offsets on a virtual address space. :param iteration: controls the inner index counter :param spacing: sets a spacing between the arrays, default is 0 All array variables (non scalars) are laid out linearly starting from 0. An optional spacing can be set. The accesses are based on this layout. The iteration 0 is the first iteration. All loops are mapped to this linear iteration space. Accesses to scalars are ignored. Returned are load and store byte-offset pairs for each iteration. """ global_load_offsets = [] global_store_offsets = [] if isinstance(iteration, range): iteration = numpy.arange(iteration.start, iteration.stop, iteration.step, dtype='O') else: if not isinstance(iteration, collections.Sequence): iteration = [iteration] iteration = numpy.array(iteration, dtype='O') # loop indices based on iteration # unwind global iteration count into loop counters: base_loop_counters = self.global_iterator_to_indices() total_length = self.iteration_length() assert iteration.max() < self.subs_consts(total_length), \ "Iterations go beyond what is possible in the original code ({} vs {}). " \ "One common reason, is that the iteration length are unrealistically small.".format( iteration.max(), self.subs_consts(total_length)) # Get sizes of arrays and base offsets for each array var_sizes = self.array_sizes(in_bytes=True, subs_consts=True) base_offsets = {} base = 0 # Always arrange arrays in alphabetical order in memory, for reproducibility for var_name, var_size in sorted(var_sizes.items(), key=lambda v: v[0]): base_offsets[var_name] = base array_total_size = self.subs_consts(var_size + spacing) # Add bytes to align by 64 byte (typical cacheline size): array_total_size = ((int(array_total_size) + 63) & ~63) base += array_total_size # Gather all read and write accesses to the array: for var_name, var_size in var_sizes.items(): element_size = self.datatypes_size[self.variables[var_name][0]] for r in self.sources.get(var_name, []): offset_expr = self.access_to_sympy(var_name, r) # Ignore accesses that always go to the same location (constant offsets) if not any([s in base_loop_counters.keys() for s in offset_expr.free_symbols]): continue offset = force_iterable(sympy.lambdify( base_loop_counters.keys(), self.subs_consts( offset_expr*element_size + base_offsets[var_name]), numpy)) # TODO possibly differentiate between index order global_load_offsets.append(offset) for w in self.destinations.get(var_name, []): offset_expr = self.access_to_sympy(var_name, w) # Ignore accesses that always go to the same location (constant offsets) if not any([s in base_loop_counters.keys() for s in offset_expr.free_symbols]): continue offset = force_iterable(sympy.lambdify( base_loop_counters.keys(), self.subs_consts( offset_expr*element_size + base_offsets[var_name]), numpy)) # TODO possibly differentiate between index order global_store_offsets.append(offset) # TODO take element sizes into account, return in bytes # Generate numpy.array for each counter counter_per_it = [v(iteration) for v in base_loop_counters.values()] # Old and slow - left for reference ## Data access as they appear with iteration order #return zip_longest(zip(*[o(*counter_per_it) for o in global_load_offsets]), # zip(*[o(*counter_per_it) for o in global_store_offsets]), # fillvalue=None) # Data access as they appear with iteration order load_offsets = [] for o in global_load_offsets: load_offsets.append(o(*counter_per_it)) # Convert to numpy ndarray and transpose to get offsets per iterations load_offsets = numpy.asarray(load_offsets).T store_offsets = [] for o in global_store_offsets: store_offsets.append(o(*counter_per_it)) store_offsets = numpy.asarray(store_offsets).T # Combine loads and stores store_width = store_offsets.shape[1] if len(store_offsets.shape) > 1 else 0 dtype = [('load', load_offsets.dtype, (load_offsets.shape[1],)), ('store', store_offsets.dtype, (store_width,))] offsets = numpy.empty(max(load_offsets.shape[0], store_offsets.shape[0]), dtype=dtype) offsets['load'] = load_offsets offsets['store'] = store_offsets return offsets
Consecutive bytes written out per high - level iterations ( as counted by loop stack ).
def bytes_per_iteration(self): """ Consecutive bytes written out per high-level iterations (as counted by loop stack). Is used to compute number of iterations per cacheline. """ # TODO Find longst consecutive writes to any variable and use as basis var_name = list(self.destinations)[0] var_type = self.variables[var_name][0] # FIXME this is correct most of the time, but not guaranteed: # Multiplying datatype size with step increment of inner-most loop return self.datatypes_size[var_type] * self._loop_stack[-1][3]
Print kernel information in human readble format.
def print_kernel_info(self, output_file=sys.stdout): """Print kernel information in human readble format.""" table = (' idx | min max step\n' + '---------+---------------------------------\n') for l in self._loop_stack: table += '{:>8} | {!r:>10} {!r:>10} {!r:>10}\n'.format(*l) print(prefix_indent('loop stack: ', table), file=output_file) table = (' name | offsets ...\n' + '---------+------------...\n') for name, offsets in list(self.sources.items()): prefix = '{:>8} | '.format(name) right_side = '\n'.join(['{!r:}'.format(o) for o in offsets]) table += prefix_indent(prefix, right_side, later_prefix=' | ') print(prefix_indent('data sources: ', table), file=output_file) table = (' name | offsets ...\n' + '---------+------------...\n') for name, offsets in list(self.destinations.items()): prefix = '{:>8} | '.format(name) right_side = '\n'.join(['{!r:}'.format(o) for o in offsets]) table += prefix_indent(prefix, right_side, later_prefix=' | ') print(prefix_indent('data destinations: ', table), file=output_file) table = (' op | count \n' + '----+-------\n') for op, count in list(self._flops.items()): table += '{:>3} | {:>4}\n'.format(op, count) table += ' =======\n' table += ' {:>4}'.format(sum(self._flops.values())) print(prefix_indent('FLOPs: ', table), file=output_file)
Print variables information in human readble format.
def print_variables_info(self, output_file=sys.stdout): """Print variables information in human readble format.""" table = (' name | type size \n' + '---------+-------------------------\n') for name, var_info in list(self.variables.items()): table += '{:>8} | {:>6} {!s:<10}\n'.format(name, var_info[0], var_info[1]) print(prefix_indent('variables: ', table), file=output_file)
Print constants information in human readble format.
def print_constants_info(self, output_file=sys.stdout): """Print constants information in human readble format.""" table = (' name | value \n' + '---------+-----------\n') for name, value in list(self.constants.items()): table += '{!s:>8} | {:<10}\n'.format(name, value) print(prefix_indent('constants: ', table), file=output_file)
Create or open intermediate file ( may be used for caching ).
def _get_intermediate_file(self, name, machine_and_compiler_dependent=True, binary=False, fp=True): """ Create or open intermediate file (may be used for caching). Will replace files older than kernel file, machine file or kerncraft version. :param machine_and_compiler_dependent: set to False if file content does not depend on machine file or compiler settings :param fp: if False, will only return file name, not file object :paarm binary: if True, use binary mode for file access :return: (file object or file name, boolean if already existent and up-to-date) """ if self._filename: base_name = os.path.join(os.path.dirname(self._filename), '.' + os.path.basename(self._filename) + '_kerncraft') else: base_name = tempfile.mkdtemp() if not self._keep_intermediates: # Remove directory and all content up on program exit atexit.register(shutil.rmtree, base_name) if machine_and_compiler_dependent: compiler, compiler_args = self._machine.get_compiler() compiler_args = '_'.join(compiler_args).replace('/', '') base_name += '/{}/{}/{}/'.format( self._machine.get_identifier(), compiler, compiler_args) # Create dirs recursively os.makedirs(base_name, exist_ok=True) # Build actual file path file_path = os.path.join(base_name, name) already_exists = False # Check if file exists and is still fresh if os.path.exists(file_path): file_modified = datetime.utcfromtimestamp(os.stat(file_path).st_mtime) if (file_modified < self._machine.get_last_modified_datetime() or file_modified < kerncraft.get_last_modified_datetime() or (self._filename and file_modified < datetime.utcfromtimestamp(os.stat(self._filename).st_mtime))): os.remove(file_path) else: already_exists = True if fp: if already_exists: mode = 'r+' else: mode = 'w' if binary: mode += 'b' f = open(file_path, mode) return f, already_exists else: return reduce_path(file_path), already_exists
Print source code of kernel.
def print_kernel_code(self, output_file=sys.stdout): """Print source code of kernel.""" print(self.kernel_code, file=output_file)
Convert mathematical expressions to a sympy representation.
def conv_ast_to_sym(self, math_ast): """ Convert mathematical expressions to a sympy representation. May only contain paranthesis, addition, subtraction and multiplication from AST. """ if type(math_ast) is c_ast.ID: return symbol_pos_int(math_ast.name) elif type(math_ast) is c_ast.Constant: return sympy.Integer(math_ast.value) else: # elif type(dim) is c_ast.BinaryOp: op = { '*': operator.mul, '+': operator.add, '-': operator.sub } return op[math_ast.op]( self.conv_ast_to_sym(math_ast.left), self.conv_ast_to_sym(math_ast.right))
Return a tuple of offsets of an ArrayRef object in all dimensions.
def _get_offsets(self, aref, dim=0): """ Return a tuple of offsets of an ArrayRef object in all dimensions. The index order is right to left (c-code order). e.g. c[i+1][j-2] -> (-2, +1) If aref is actually a c_ast.ID, None will be returned. """ if isinstance(aref, c_ast.ID): return None # Check for restrictions assert type(aref.name) in [c_ast.ArrayRef, c_ast.ID], \ "array references must only be used with variables or other array references" assert type(aref.subscript) in [c_ast.ID, c_ast.Constant, c_ast.BinaryOp], \ 'array subscript must only contain variables or binary operations' # Convert subscript to sympy and append idxs = [self.conv_ast_to_sym(aref.subscript)] # Check for more indices (multi-dimensional access) if type(aref.name) is c_ast.ArrayRef: idxs += self._get_offsets(aref.name, dim=dim+1) # Reverse to preserver order (the subscripts in the AST are traversed backwards) if dim == 0: idxs.reverse() return tuple(idxs)
Return base name of ArrayRef object.
def _get_basename(cls, aref): """ Return base name of ArrayRef object. e.g. c[i+1][j-2] -> 'c' """ if isinstance(aref.name, c_ast.ArrayRef): return cls._get_basename(aref.name) elif isinstance(aref.name, str): return aref.name else: return aref.name.name
Return index type used in loop nest.
def get_index_type(self, loop_nest=None): """ Return index type used in loop nest. If index type between loops differ, an exception is raised. """ if loop_nest is None: loop_nest = self.get_kernel_loop_nest() if type(loop_nest) is c_ast.For: loop_nest = [loop_nest] index_types = (None, None) for s in loop_nest: if type(s) is c_ast.For: if type(s.stmt) in [c_ast.For, c_ast.Compound]: other = self.get_index_type(loop_nest=s.stmt) else: other = None index_types = (s.init.decls[0].type.type.names, other) break if index_types[0] == index_types[1] or index_types[1] is None: return index_types[0] else: raise ValueError("Loop indices must have same type, found {}.".format(index_types))
Generate constants declarations
def _build_const_declartions(self, with_init=True): """ Generate constants declarations :return: list of declarations """ decls = [] # Use type as provided by user in loop indices index_type = self.get_index_type() i = 2 # subscript for cli input, 1 is reserved for repeat for k in self.constants: # const long long N = strtoul(argv[2]) # with increasing N and 1 # TODO change subscript of argv depending on constant count type_decl = c_ast.TypeDecl(k.name, ['const'], c_ast.IdentifierType(index_type)) init = None if with_init: init = c_ast.FuncCall( c_ast.ID('atoi'), c_ast.ExprList([c_ast.ArrayRef(c_ast.ID('argv'), c_ast.Constant('int', str(i)))])) i += 1 decls.append(c_ast.Decl( k.name, ['const'], [], [], type_decl, init, None)) return decls
Return array declarations.
def get_array_declarations(self): """Return array declarations.""" return [d for d in self.kernel_ast.block_items if type(d) is c_ast.Decl and type(d.type) is c_ast.ArrayDecl]
Return kernel loop nest including any preceding pragmas and following swaps.
def get_kernel_loop_nest(self): """Return kernel loop nest including any preceding pragmas and following swaps.""" loop_nest = [s for s in self.kernel_ast.block_items if type(s) in [c_ast.For, c_ast.Pragma, c_ast.FuncCall]] assert len(loop_nest) >= 1, "Found to few for statements in kernel" return loop_nest
Generate declaration statements for arrays.
def _build_array_declarations(self, with_init=True): """ Generate declaration statements for arrays. Also transforming multi-dim to 1d arrays and initializing with malloc. :param with_init: ommit malloc initialization :return: list of declarations nodes, dictionary of array names and original dimensions """ # copy array declarations from from kernel ast array_declarations = deepcopy(self.get_array_declarations()) array_dict = [] for d in array_declarations: # We need to transform array_dict.append(transform_multidim_to_1d_decl(d)) transform_array_decl_to_malloc(d, with_init=with_init) return array_declarations, dict(array_dict)
Return inner most for loop in loop nest
def _find_inner_most_loop(self, loop_nest): """Return inner most for loop in loop nest""" r = None for s in loop_nest: if type(s) is c_ast.For: return self._find_inner_most_loop(s) or s else: r = r or self._find_inner_most_loop(s) return r
Generate initialization statements for arrays.
def _build_array_initializations(self, array_dimensions): """ Generate initialization statements for arrays. :param array_dimensions: dictionary of array dimensions :return: list of nodes """ kernel = deepcopy(deepcopy(self.get_kernel_loop_nest())) # traverse to the inner most for loop: inner_most = self._find_inner_most_loop(kernel) orig_inner_stmt = inner_most.stmt inner_most.stmt = c_ast.Compound([]) rand_float_str = str(random.uniform(1.0, 0.1)) # find all array references in original orig_inner_stmt for aref in find_node_type(orig_inner_stmt, c_ast.ArrayRef): # transform to 1d references transform_multidim_to_1d_ref(aref, array_dimensions) # build static assignments and inject into inner_most.stmt inner_most.stmt.block_items.append(c_ast.Assignment( '=', aref, c_ast.Constant('float', rand_float_str))) return kernel
Generate false if branch with dummy calls
def _build_dummy_calls(self): """ Generate false if branch with dummy calls Requires kerncraft.h to be included, which defines dummy(...) and var_false. :return: dummy statement """ # Make sure nothing gets removed by inserting dummy calls dummy_calls = [] for d in self.kernel_ast.block_items: # Only consider toplevel declarations from kernel ast if type(d) is not c_ast.Decl: continue if type(d.type) is c_ast.ArrayDecl: dummy_calls.append(c_ast.FuncCall( c_ast.ID('dummy'), c_ast.ExprList([c_ast.ID(d.name)]))) else: dummy_calls.append(c_ast.FuncCall( c_ast.ID('dummy'), c_ast.ExprList([c_ast.UnaryOp('&', c_ast.ID(d.name))]))) dummy_stmt = c_ast.If( cond=c_ast.ID('var_false'), iftrue=c_ast.Compound(dummy_calls), iffalse=None) return dummy_stmt
Build and return kernel function declaration
def _build_kernel_function_declaration(self, name='kernel'): """Build and return kernel function declaration""" array_declarations, array_dimensions = self._build_array_declarations(with_init=False) scalar_declarations = self._build_scalar_declarations(with_init=False) const_declarations = self._build_const_declartions(with_init=False) return c_ast.FuncDecl(args=c_ast.ParamList(params=array_declarations + scalar_declarations + const_declarations), type=c_ast.TypeDecl(declname=name, quals=[], type=c_ast.IdentifierType(names=['void'])))
Build and return scalar variable declarations
def _build_scalar_declarations(self, with_init=True): """Build and return scalar variable declarations""" # copy scalar declarations from from kernel ast scalar_declarations = [deepcopy(d) for d in self.kernel_ast.block_items if type(d) is c_ast.Decl and type(d.type) is c_ast.TypeDecl] # add init values to declarations if with_init: random.seed(2342) # we want reproducible random numbers for d in scalar_declarations: if d.type.type.names[0] in ['double', 'float']: d.init = c_ast.Constant('float', str(random.uniform(1.0, 0.1))) elif d.type.type.names[0] in ['int', 'long', 'long long', 'unsigned int', 'unsigned long', 'unsigned long long']: d.init = c_ast.Constant('int', 2) return scalar_declarations
Generate and return compilable source code with kernel function from AST.
def get_kernel_code(self, openmp=False, as_filename=False, name='kernel'): """ Generate and return compilable source code with kernel function from AST. :param openmp: if true, OpenMP code will be generated :param as_filename: if true, will save to file and return filename :param name: name of kernel function """ assert self.kernel_ast is not None, "AST does not exist, this could be due to running " \ "based on a kernel description rather than code." file_name = 'kernel' if openmp: file_name += '-omp' file_name += '.c' fp, already_available = self._get_intermediate_file( file_name, machine_and_compiler_dependent=False) # Use already cached version if already_available: code = fp.read() else: array_declarations, array_dimensions = self._build_array_declarations() # Prepare actual kernel loop nest if openmp: # with OpenMP code kernel = deepcopy(self.get_kernel_loop_nest()) # find all array references in kernel for aref in find_node_type(kernel, c_ast.ArrayRef): # transform to 1d references transform_multidim_to_1d_ref(aref, array_dimensions) omp_pragmas = [p for p in find_node_type(kernel, c_ast.Pragma) if 'omp' in p.string] # TODO if omp parallel was found, remove it (also replace "parallel for" -> "for") # if no omp for pragmas are present, insert suitable ones if not omp_pragmas: kernel.insert(0, c_ast.Pragma("omp for")) # otherwise do not change anything else: # with original code kernel = deepcopy(self.get_kernel_loop_nest()) # find all array references in kernel for aref in find_node_type(kernel, c_ast.ArrayRef): # transform to 1d references transform_multidim_to_1d_ref(aref, array_dimensions) function_ast = c_ast.FuncDef(decl=c_ast.Decl( name=name, type=self._build_kernel_function_declaration(name=name), quals=[], storage=[], funcspec=[], init=None, bitsize=None), body=c_ast.Compound(block_items=kernel), param_decls=None) # Generate code code = CGenerator().visit(function_ast) # Insert missing #includes from template to top of code code = '#include "kerncraft.h"\n\n' + code # Store to file fp.write(code) fp.close() if as_filename: return fp.name else: return code
Generate and return kernel call ast.
def _build_kernel_call(self, name='kernel'): """Generate and return kernel call ast.""" return c_ast.FuncCall(name=c_ast.ID(name=name), args=c_ast.ExprList(exprs=[ c_ast.ID(name=d.name) for d in ( self._build_array_declarations()[0] + self._build_scalar_declarations() + self._build_const_declartions())]))
Generate and return compilable source code from AST.
def get_main_code(self, as_filename=False, kernel_function_name='kernel'): """ Generate and return compilable source code from AST. """ # TODO produce nicer code, including help text and other "comfort features". assert self.kernel_ast is not None, "AST does not exist, this could be due to running " \ "based on a kernel description rather than code." fp, already_available = self._get_intermediate_file('main.c', machine_and_compiler_dependent=False) # Use already cached version if already_available: code = fp.read() else: parser = CParser() template_code = self.CODE_TEMPLATE template_ast = parser.parse(clean_code(template_code, macros=True, comments=True, pragmas=False)) ast = deepcopy(template_ast) # Define and replace DECLARE_CONSTS replace_id(ast, "DECLARE_CONSTS", self._build_const_declartions(with_init=True)) # Define and replace DECLARE_ARRAYS array_declarations, array_dimensions = self._build_array_declarations() replace_id(ast, "DECLARE_ARRAYS", array_declarations) # Define and replace DECLARE_INIT_SCALARS replace_id(ast, "DECLARE_INIT_SCALARS", self._build_scalar_declarations()) # Define and replace DUMMY_CALLS replace_id(ast, "DUMMY_CALLS", self._build_dummy_calls()) # Define and replace KERNEL_DECL ast.ext.insert(0, self._build_kernel_function_declaration( name=kernel_function_name)) # Define and replace KERNEL_CALL replace_id(ast, "KERNEL_CALL", self._build_kernel_call()) # Define and replace INIT_ARRAYS based on previously generated kernel replace_id(ast, "INIT_ARRAYS", self._build_array_initializations(array_dimensions)) # Generate code code = CGenerator().visit(ast) # Insert missing #includes from template to top of code code = '\n'.join([l for l in template_code.split('\n') if l.startswith("#include")]) + \ '\n\n' + code # Store to file fp.write(code) fp.close() if as_filename: return fp.name else: return code
Assemble * in_filename * assembly into * out_filename * object.
def assemble_to_object(self, in_filename, verbose=False): """ Assemble *in_filename* assembly into *out_filename* object. If *iaca_marked* is set to true, markers are inserted around the block with most packed instructions or (if no packed instr. were found) the largest block and modified file is saved to *in_file*. *asm_block* controls how the to-be-marked block is chosen. "auto" (default) results in the largest block, "manual" results in interactive and a number in the according block. *pointer_increment* is the number of bytes the pointer is incremented after the loop or - 'auto': automatic detection, RuntimeError is raised in case of failure - 'auto_with_manual_fallback': automatic detection, fallback to manual input - 'manual': prompt user Returns two-tuple (filepointer, filename) to temp binary file. """ # Build file name file_base_name = os.path.splitext(os.path.basename(in_filename))[0] out_filename, already_exists = self._get_intermediate_file(file_base_name + '.o', binary=True, fp=False) if already_exists: # Do not use caching, because pointer_increment or asm_block selection may be different pass compiler, compiler_args = self._machine.get_compiler() # Compile to object file compiler_args.append('-c') cmd = [compiler] + [ in_filename] + \ compiler_args + ['-o', out_filename] if verbose: print('Executing (assemble_to_object): ', ' '.join(cmd)) try: # Assemble all to a binary subprocess.check_output(cmd) except subprocess.CalledProcessError as e: print("Assembly failed:", e, file=sys.stderr) sys.exit(1) return out_filename
Compile source ( from as_code ( type_ )) to assembly or object and return ( fileptr filename ).
def compile_kernel(self, openmp=False, assembly=False, verbose=False): """ Compile source (from as_code(type_)) to assembly or object and return (fileptr, filename). Output can be used with Kernel.assemble() """ compiler, compiler_args = self._machine.get_compiler() in_filename = self.get_kernel_code(openmp=openmp, as_filename=True) if assembly: compiler_args += ['-S'] suffix = '.s' else: suffix = '.o' out_filename, already_exists = self._get_intermediate_file( os.path.splitext(os.path.basename(in_filename))[0]+suffix, binary=not assembly, fp=False) if already_exists: if verbose: print('Executing (compile_kernel): ', 'using cached', out_filename) return out_filename compiler_args += ['-std=c99'] cmd = ([compiler] + [in_filename, '-c', '-I'+reduce_path(os.path.abspath(os.path.dirname( os.path.realpath(__file__)))+'/headers/'), '-o', out_filename] + compiler_args) if verbose: print('Executing (compile_kernel): ', ' '.join(cmd)) try: subprocess.check_output(cmd) except subprocess.CalledProcessError as e: print("Compilation failed:", e, file=sys.stderr) sys.exit(1) # FIXME TODO FIXME TODO FIXME TODO # Hacky workaround for icc issue (icc may issue vkmovb instructions with AVX512, which are # invalid and should be kmovb): if compiler == 'icc' and assembly: with open(out_filename, 'r+') as f: assembly = f.read() f.seek(0) f.write(assembly.replace('vkmovb', 'kmovb')) f.truncate() # FIXME TODO FIXME TODO FIXME TODO # Let's return the out_file name return out_filename
Run an IACA analysis and return its outcome.
def iaca_analysis(self, micro_architecture, asm_block='auto', pointer_increment='auto_with_manual_fallback', verbose=False): """ Run an IACA analysis and return its outcome. *asm_block* controls how the to-be-marked block is chosen. "auto" (default) results in the largest block, "manual" results in interactive and a number in the according block. *pointer_increment* is the number of bytes the pointer is incremented after the loop or - 'auto': automatic detection, RuntimeError is raised in case of failure - 'auto_with_manual_fallback': automatic detection, fallback to manual input - 'manual': prompt user """ asm_filename = self.compile_kernel(assembly=True, verbose=verbose) asm_marked_filename = os.path.splitext(asm_filename)[0]+'-iaca.s' with open(asm_filename, 'r') as in_file, open(asm_marked_filename, 'w') as out_file: self.asm_block = iaca.iaca_instrumentation( in_file, out_file, block_selection=asm_block, pointer_increment=pointer_increment) obj_name = self.assemble_to_object(asm_marked_filename, verbose=verbose) return iaca.iaca_analyse_instrumented_binary(obj_name, micro_architecture), self.asm_block
Compile source to executable with likwid capabilities and return the executable name.
def build_executable(self, lflags=None, verbose=False, openmp=False): """Compile source to executable with likwid capabilities and return the executable name.""" compiler, compiler_args = self._machine.get_compiler() kernel_obj_filename = self.compile_kernel(openmp=openmp, verbose=verbose) out_filename, already_exists = self._get_intermediate_file( os.path.splitext(os.path.basename(kernel_obj_filename))[0], binary=True, fp=False) if not already_exists: main_source_filename = self.get_main_code(as_filename=True) if not (('LIKWID_INCLUDE' in os.environ or 'LIKWID_INC' in os.environ) and 'LIKWID_LIB' in os.environ): print('Could not find LIKWID_INCLUDE (e.g., "-I/app/likwid/4.1.2/include") and ' 'LIKWID_LIB (e.g., "-L/apps/likwid/4.1.2/lib") environment variables', file=sys.stderr) sys.exit(1) compiler_args += [ '-std=c99', '-I'+reduce_path(os.path.abspath(os.path.dirname( os.path.realpath(__file__)))+'/headers/'), os.environ.get('LIKWID_INCLUDE', ''), os.environ.get('LIKWID_INC', ''), '-llikwid'] # This is a special case for unittesting if os.environ.get('LIKWID_LIB') == '': compiler_args = compiler_args[:-1] if lflags is None: lflags = [] lflags += os.environ['LIKWID_LIB'].split(' ') + ['-pthread'] compiler_args += os.environ['LIKWID_LIB'].split(' ') + ['-pthread'] infiles = [reduce_path(os.path.abspath(os.path.dirname( os.path.realpath(__file__)))+'/headers/dummy.c'), kernel_obj_filename, main_source_filename] cmd = [compiler] + infiles + compiler_args + ['-o', out_filename] # remove empty arguments cmd = list(filter(bool, cmd)) if verbose: print('Executing (build_executable): ', ' '.join(cmd)) try: subprocess.check_output(cmd) except subprocess.CalledProcessError as e: print("Build failed:", e, file=sys.stderr) sys.exit(1) else: if verbose: print('Executing (build_executable): ', 'using cached', out_filename) return out_filename
Convert any string to a sympy object or None.
def string_to_sympy(cls, s): """Convert any string to a sympy object or None.""" if isinstance(s, int): return sympy.Integer(s) elif isinstance(s, list): return tuple([cls.string_to_sympy(e) for e in s]) elif s is None: return None else: # Step 1 build expression with the whole alphabet redefined: local_dict = {c: symbol_pos_int(c) for c in s if c in string.ascii_letters} # TODO find nicer solution for N and other pre-mapped letters preliminary_expr = parse_expr(s, local_dict=local_dict) # Replace all free symbols with positive integer versions: local_dict.update( {s.name: symbol_pos_int(s.name) for s in preliminary_expr.free_symbols}) return parse_expr(s, local_dict=local_dict)
Return identifier which is either the machine file name or sha256 checksum of data.
def get_identifier(self): """Return identifier which is either the machine file name or sha256 checksum of data.""" if self._path: return os.path.basename(self._path) else: return hashlib.sha256(hashlib.sha256(repr(self._data).encode())).hexdigest()
Return datetime object of modified time of machine file. Return now if not a file.
def get_last_modified_datetime(self): """Return datetime object of modified time of machine file. Return now if not a file.""" if self._path: statbuf = os.stat(self._path) return datetime.utcfromtimestamp(statbuf.st_mtime) else: return datetime.now()
Return a cachesim. CacheSimulator object based on the machine description.
def get_cachesim(self, cores=1): """ Return a cachesim.CacheSimulator object based on the machine description. :param cores: core count (default: 1) """ cache_dict = {} for c in self['memory hierarchy']: # Skip main memory if 'cache per group' not in c: continue cache_dict[c['level']] = deepcopy(c['cache per group']) # Scale size of shared caches according to cores if c['cores per group'] > 1: cache_dict[c['level']]['sets'] //= cores cs, caches, mem = cachesim.CacheSimulator.from_dict(cache_dict) return cs
Return best fitting bandwidth according to number of threads read and write streams.
def get_bandwidth(self, cache_level, read_streams, write_streams, threads_per_core, cores=None): """ Return best fitting bandwidth according to number of threads, read and write streams. :param cache_level: integer of cache (0 is L1, 1 is L2 ...) :param read_streams: number of read streams expected :param write_streams: number of write streams expected :param threads_per_core: number of threads that are run on each core :param cores: if not given, will choose maximum bandwidth for single NUMA domain """ # try to find best fitting kernel (closest to read/write ratio): # write allocate has to be handled in kernel information (all writes are also reads) # TODO support for non-write-allocate architectures try: target_ratio = read_streams/write_streams except ZeroDivisionError: target_ratio = float('inf') measurement_kernel = 'load' measurement_kernel_info = self['benchmarks']['kernels'][measurement_kernel] measurement_kernel_ratio = float('inf') for kernel_name, kernel_info in sorted(self['benchmarks']['kernels'].items()): try: kernel_ratio = ((kernel_info['read streams']['streams'] + kernel_info['write streams']['streams'] - kernel_info['read+write streams']['streams']) / kernel_info['write streams']['streams']) except ZeroDivisionError: kernel_ratio = float('inf') if abs(kernel_ratio - target_ratio) < abs(measurement_kernel_ratio - target_ratio): measurement_kernel = kernel_name measurement_kernel_info = kernel_info measurement_kernel_ratio = kernel_ratio # choose smt, and then use max/saturation bw bw_level = self['memory hierarchy'][cache_level]['level'] bw_measurements = \ self['benchmarks']['measurements'][bw_level][threads_per_core] assert threads_per_core == bw_measurements['threads per core'], \ 'malformed measurement dictionary in machine file.' if cores is not None: # Used by Roofline model run_index = bw_measurements['cores'].index(cores) bw = bw_measurements['results'][measurement_kernel][run_index] else: # Used by ECM model # Choose maximum number of cores to get bandwidth for max_cores = min(self['memory hierarchy'][cache_level]['cores per group'], self['cores per NUMA domain']) bw = max(bw_measurements['results'][measurement_kernel][:max_cores]) # Correct bandwidth due to miss-measurement of write allocation # TODO support non-temporal stores and non-write-allocate architectures if cache_level == 0: # L1 does not have write-allocate, so everything is measured correctly factor = 1.0 else: factor = (float(measurement_kernel_info['read streams']['bytes']) + 2.0*float(measurement_kernel_info['write streams']['bytes']) - float(measurement_kernel_info['read+write streams']['bytes'])) / \ (float(measurement_kernel_info['read streams']['bytes']) + float(measurement_kernel_info['write streams']['bytes'])) bw = bw * factor return bw, measurement_kernel
Return tuple of compiler and compiler flags.
def get_compiler(self, compiler=None, flags=None): """ Return tuple of compiler and compiler flags. Selects compiler and flags from machine description file, commandline arguments or call arguements. """ if self._args: compiler = compiler or self._args.compiler flags = flags or self._args.compiler_flags if compiler is None: # Select first available compiler in machine description file's compiler dict for c in self['compiler'].keys(): # Making sure compiler is available: if find_executable(c) is not None: compiler = c break else: raise RuntimeError("No compiler ({}) was found. Add different one in machine file, " "via --compiler argument or make sure it will be found in " "$PATH.".format(list(self['compiler'].keys()))) if flags is None: # Select from machine description file flags = self['compiler'].get(compiler, '') return compiler, flags.split(' ')
Parse events in machine description to tuple representation used in Benchmark module.
def parse_perfctr_event(perfctr): """ Parse events in machine description to tuple representation used in Benchmark module. Examples: >>> parse_perfctr_event('PERF_EVENT:REG[0-3]') ('PERF_EVENT', 'REG[0-3]') >>> parse_perfctr_event('PERF_EVENT:REG[0-3]:STAY:FOO=23:BAR=0x23') ('PERF_EVENT', 'REG[0-3]', {'STAY': None, 'FOO': 23, 'BAR': 35}) """ split_perfctr = perfctr.split(':') assert len(split_perfctr) >= 2, "Atleast one colon (:) is required in the event name" event_tuple = split_perfctr[:2] parameters = {} for p in split_perfctr[2:]: if '=' in p: k, v = p.split('=') if v.startswith('0x'): parameters[k] = int(v, 16) else: parameters[k] = int(v) else: parameters[p] = None event_tuple.append(parameters) return tuple(event_tuple)
Return ( sympy expressions event names and symbols dict ) from performance metric str.
def parse_perfmetric(metric): """Return (sympy expressions, event names and symbols dict) from performance metric str.""" # Find all perfs counter references perfcounters = re.findall(r'[A-Z0-9_]+:[A-Z0-9\[\]|\-]+(?::[A-Za-z0-9\-_=]+)*', metric) # Build a temporary metric, with parser-friendly Symbol names temp_metric = metric temp_pc_names = {"SYM{}".format(re.sub("[\[\]\-|=:]", "_", pc)): pc for i, pc in enumerate(perfcounters)} for var_name, pc in temp_pc_names.items(): temp_metric = temp_metric.replace(pc, var_name) # Parse temporary expression expr = parse_expr(temp_metric) # Rename symbols to originals for s in expr.free_symbols: if s.name in temp_pc_names: s.name = temp_pc_names[str(s)] events = {s: MachineModel.parse_perfctr_event(s.name) for s in expr.free_symbols if s.name in perfcounters} return expr, events
Enforce that no ranges overlap in internal storage.
def _enforce_no_overlap(self, start_at=0): """Enforce that no ranges overlap in internal storage.""" i = start_at while i+1 < len(self.data): if self.data[i][1] >= self.data[i+1][0]: # beginning of i+1-th range is contained in i-th range if self.data[i][1] < self.data[i+1][1]: # i+1-th range is longer, thus enlarge i-th range self.data[i][1] = self.data[i+1][1] # removed contained range del self.data[i+1] i += 1
Return local folder path of header files.
def get_header_path() -> str: """Return local folder path of header files.""" import os return os.path.abspath(os.path.dirname(os.path.realpath(__file__))) + '/headers/'
Align iteration with cacheline boundary.
def _align_iteration_with_cl_boundary(self, iteration, subtract=True): """Align iteration with cacheline boundary.""" # FIXME handle multiple datatypes element_size = self.kernel.datatypes_size[self.kernel.datatype] cacheline_size = self.machine['cacheline size'] elements_per_cacheline = int(cacheline_size // element_size) # Gathering some loop information: inner_loop = list(self.kernel.get_loop_stack(subs_consts=True))[-1] inner_increment = inner_loop['increment'] # do this by aligning either writes (preferred) or reads # Assumption: writes (and reads) increase linearly o = self.kernel.compile_global_offsets(iteration=iteration)[0] if len(o[1]): # we have a write to work with: first_offset = min(o[1]) else: # we use reads first_offset = min(o[0]) diff = first_offset - \ (int(first_offset) >> self.csim.first_level.cl_bits << self.csim.first_level.cl_bits) if diff == 0: return iteration elif subtract: return iteration - (diff // element_size) // inner_increment else: return iteration + (elements_per_cacheline - diff // element_size) // inner_increment
Return a list with number of loaded cache lines per memory hierarchy level.
def get_loads(self): """Return a list with number of loaded cache lines per memory hierarchy level.""" return [self.stats[cache_level]['LOAD_count'] / self.first_dim_factor for cache_level in range(len(self.machine['memory hierarchy']))]
Return a list with number of hit cache lines per memory hierarchy level.
def get_hits(self): """Return a list with number of hit cache lines per memory hierarchy level.""" return [self.stats[cache_level]['HIT_count']/self.first_dim_factor for cache_level in range(len(self.machine['memory hierarchy']))]
Return a list with number of missed cache lines per memory hierarchy level.
def get_misses(self): """Return a list with number of missed cache lines per memory hierarchy level.""" return [self.stats[cache_level]['MISS_count']/self.first_dim_factor for cache_level in range(len(self.machine['memory hierarchy']))]
Return a list with number of stored cache lines per memory hierarchy level.
def get_stores(self): """Return a list with number of stored cache lines per memory hierarchy level.""" return [self.stats[cache_level]['STORE_count']/self.first_dim_factor for cache_level in range(len(self.machine['memory hierarchy']))]
Return a list with number of evicted cache lines per memory hierarchy level.
def get_evicts(self): """Return a list with number of evicted cache lines per memory hierarchy level.""" return [self.stats[cache_level]['EVICT_count']/self.first_dim_factor for cache_level in range(len(self.machine['memory hierarchy']))]
Return verbose information about the predictor.
def get_infos(self): """Return verbose information about the predictor.""" first_dim_factor = self.first_dim_factor infos = {'memory hierarchy': [], 'cache stats': self.stats, 'cachelines in stats': first_dim_factor} for cache_level, cache_info in list(enumerate(self.machine['memory hierarchy'])): infos['memory hierarchy'].append({ 'index': len(infos['memory hierarchy']), 'level': '{}'.format(cache_info['level']), 'total loads': self.stats[cache_level]['LOAD_byte']/first_dim_factor, 'total misses': self.stats[cache_level]['MISS_byte']/first_dim_factor, 'total hits': self.stats[cache_level]['HIT_byte']/first_dim_factor, 'total stores': self.stats[cache_level]['STORE_byte']/first_dim_factor, 'total evicts': self.stats[cache_level]['EVICT_byte']/first_dim_factor, 'total lines load': self.stats[cache_level]['LOAD_count']/first_dim_factor, 'total lines misses': self.stats[cache_level]['MISS_count']/first_dim_factor, 'total lines hits': self.stats[cache_level]['HIT_count']/first_dim_factor, 'total lines stores': self.stats[cache_level]['STORE_count']/first_dim_factor, 'total lines evicts': self.stats[cache_level]['EVICT_count']/first_dim_factor, 'cycles': None}) return infos
* size * is given in kilo bytes
def measure_bw(type_, total_size, threads_per_core, max_threads_per_core, cores_per_socket, sockets): """*size* is given in kilo bytes""" groups = [] for s in range(sockets): groups += [ '-w', 'S' + str(s) + ':' + str(total_size) + 'kB:' + str(threads_per_core * cores_per_socket) + ':1:' + str(int(max_threads_per_core / threads_per_core))] # for older likwid versions add ['-g', str(sockets), '-i', str(iterations)] to cmd cmd = ['likwid-bench', '-t', type_] + groups sys.stderr.write(' '.join(cmd)) output = subprocess.Popen(cmd, stdout=subprocess.PIPE).communicate()[0].decode('utf-8') if not output: print(' '.join(cmd) + ' returned no output, possibly wrong version installed ' '(requires 4.0 or later)', file=sys.stderr) sys.exit(1) bw = float(get_match_or_break(r'^MByte/s:\s+([0-9]+(?:\.[0-9]+)?)\s*$', output)[0]) print(' ', PrefixedUnit(bw, 'MB/s'), file=sys.stderr) return PrefixedUnit(bw, 'MB/s')
Fix environment variable to a value within context. Unset if value is None.
def fix_env_variable(name, value): """Fix environment variable to a value within context. Unset if value is None.""" orig = os.environ.get(name, None) if value is not None: # Set if value is not None os.environ[name] = value elif name in os.environ: # Unset if value is None del os.environ[name] try: yield finally: if orig is not None: # Restore original value os.environ[name] = orig elif name in os.environ: # Unset del os.environ[name]
Configure argument parser.
def configure_arggroup(cls, parser): """Configure argument parser.""" parser.add_argument( '--no-phenoecm', action='store_true', help='Disables the phenomenological ECM model building.') parser.add_argument( '--iterations', type=int, default=10, help='Number of outer-loop iterations (e.g. time loop) during benchmarking. ' 'Default is 10, but actual number will be adapted to at least 0.2s runtime.') parser.add_argument( '--ignore-warnings', action='store_true', help='Ignore warnings about missmatched CPU model and frequency.')
Run * cmd * with likwid - perfctr and returns result as dict.
def perfctr(self, cmd, group='MEM', code_markers=True): """ Run *cmd* with likwid-perfctr and returns result as dict. *group* may be a performance group known to likwid-perfctr or an event string. if CLI argument cores > 1, running with multi-core, otherwise single-core """ # Making sure likwid-perfctr is available: if find_executable('likwid-perfctr') is None: print("likwid-perfctr was not found. Make sure likwid is installed and found in PATH.", file=sys.stderr) sys.exit(1) # FIXME currently only single core measurements support! perf_cmd = ['likwid-perfctr', '-f', '-O', '-g', group] cpu = 'S0:0' if self._args.cores > 1: cpu += '-'+str(self._args.cores-1) # Pinned and measured on cpu perf_cmd += ['-C', cpu] # code must be marked using likwid markers perf_cmd.append('-m') perf_cmd += cmd if self.verbose > 1: print(' '.join(perf_cmd)) try: with fix_env_variable('OMP_NUM_THREADS', None): output = subprocess.check_output(perf_cmd).decode('utf-8').split('\n') except subprocess.CalledProcessError as e: print("Executing benchmark failed: {!s}".format(e), file=sys.stderr) sys.exit(1) # TODO multicore output is different and needs to be considered here! results = {} for line in output: line = line.split(',') try: # Metrics results[line[0]] = float(line[1]) except ValueError: # Would not convert to float pass except IndexError: # Not a parable line (did not contain any commas) continue try: # Event counters if line[2] == '-' or line[2] == 'nan': counter_value = 0 else: counter_value = int(line[2]) if re.fullmatch(r'[A-Z0-9_]+', line[0]) and re.fullmatch(r'[A-Z0-9]+', line[1]): results.setdefault(line[0], {}) results[line[0]][line[1]] = counter_value except (IndexError, ValueError): pass return results
Run analysis.
def analyze(self): """Run analysis.""" bench = self.kernel.build_executable(verbose=self.verbose > 1, openmp=self._args.cores > 1) element_size = self.kernel.datatypes_size[self.kernel.datatype] # Build arguments to pass to command: args = [str(s) for s in list(self.kernel.constants.values())] # Determine base runtime with 10 iterations runtime = 0.0 time_per_repetition = 2.0 / 10.0 repetitions = self.iterations // 10 mem_results = {} # TODO if cores > 1, results are for openmp run. Things might need to be changed here! while runtime < 1.5: # Interpolate to a 2.0s run if time_per_repetition != 0.0: repetitions = 2.0 // time_per_repetition else: repetitions = int(repetitions * 10) mem_results = self.perfctr([bench] + [str(repetitions)] + args, group="MEM") runtime = mem_results['Runtime (RDTSC) [s]'] time_per_repetition = runtime / float(repetitions) raw_results = [mem_results] # Base metrics for further metric computations: # An iteration is equal to one high-level code inner-most-loop iteration iterations_per_repetition = reduce( operator.mul, [self.kernel.subs_consts(max_ - min_) / self.kernel.subs_consts(step) for idx, min_, max_, step in self.kernel._loop_stack], 1) iterations_per_cacheline = (float(self.machine['cacheline size']) / self.kernel.bytes_per_iteration) cys_per_repetition = time_per_repetition * float(self.machine['clock']) # Gather remaining counters if not self.no_phenoecm: # Build events and sympy expressions for all model metrics T_OL, event_counters = self.machine.parse_perfmetric( self.machine['overlapping model']['performance counter metric']) T_data, event_dict = self.machine.parse_perfmetric( self.machine['non-overlapping model']['performance counter metric']) event_counters.update(event_dict) cache_metrics = defaultdict(dict) for i in range(len(self.machine['memory hierarchy']) - 1): cache_info = self.machine['memory hierarchy'][i] name = cache_info['level'] for k, v in cache_info['performance counter metrics'].items(): cache_metrics[name][k], event_dict = self.machine.parse_perfmetric(v) event_counters.update(event_dict) # Compile minimal runs to gather all required events minimal_runs = build_minimal_runs(list(event_counters.values())) measured_ctrs = {} for run in minimal_runs: ctrs = ','.join([eventstr(e) for e in run]) r = self.perfctr([bench] + [str(repetitions)] + args, group=ctrs) raw_results.append(r) measured_ctrs.update(r) # Match measured counters to symbols event_counter_results = {} for sym, ctr in event_counters.items(): event, regs, parameter = ctr[0], register_options(ctr[1]), ctr[2] for r in regs: if r in measured_ctrs[event]: event_counter_results[sym] = measured_ctrs[event][r] # Analytical metrics needed for futher calculation cl_size = float(self.machine['cacheline size']) total_iterations = iterations_per_repetition * repetitions total_cachelines = total_iterations / iterations_per_cacheline T_OL_result = T_OL.subs(event_counter_results) / total_cachelines cache_metric_results = defaultdict(dict) for cache, mtrcs in cache_metrics.items(): for m, e in mtrcs.items(): cache_metric_results[cache][m] = e.subs(event_counter_results) # Inter-cache transfers per CL cache_transfers_per_cl = {cache: {k: PrefixedUnit(v / total_cachelines, 'CL/CL') for k, v in d.items()} for cache, d in cache_metric_results.items()} cache_transfers_per_cl['L1']['accesses'].unit = 'LOAD/CL' # Select appropriate bandwidth mem_bw, mem_bw_kernel = self.machine.get_bandwidth( -1, # mem cache_metric_results['L3']['misses'], # load_streams cache_metric_results['L3']['evicts'], # store_streams 1) data_transfers = { # Assuming 0.5 cy / LOAD (SSE on SNB or IVB; AVX on HSW, BDW, SKL or SKX) 'T_nOL': (cache_metric_results['L1']['accesses'] / total_cachelines * 0.5), 'T_L1L2': ((cache_metric_results['L1']['misses'] + cache_metric_results['L1']['evicts']) / total_cachelines * cl_size / self.machine['memory hierarchy'][1]['non-overlap upstream throughput'][0]), 'T_L2L3': ((cache_metric_results['L2']['misses'] + cache_metric_results['L2']['evicts']) / total_cachelines * cl_size / self.machine['memory hierarchy'][2]['non-overlap upstream throughput'][0]), 'T_L3MEM': ((cache_metric_results['L3']['misses'] + cache_metric_results['L3']['evicts']) * float(self.machine['cacheline size']) / total_cachelines / mem_bw * float(self.machine['clock'])) } # Build phenomenological ECM model: ecm_model = {'T_OL': T_OL_result} ecm_model.update(data_transfers) else: event_counters = {} ecm_model = None cache_transfers_per_cl = None self.results = {'raw output': raw_results, 'ECM': ecm_model, 'data transfers': cache_transfers_per_cl, 'Runtime (per repetition) [s]': time_per_repetition, 'event counters': event_counters, 'Iterations per repetition': iterations_per_repetition, 'Iterations per cacheline': iterations_per_cacheline} # TODO make more generic to support other (and multiple) constant names self.results['Runtime (per cacheline update) [cy/CL]'] = \ (cys_per_repetition / iterations_per_repetition) * iterations_per_cacheline self.results['MEM volume (per repetition) [B]'] = \ mem_results['Memory data volume [GBytes]'] * 1e9 / repetitions self.results['Performance [MFLOP/s]'] = \ sum(self.kernel._flops.values()) / ( time_per_repetition / iterations_per_repetition) / 1e6 if 'Memory bandwidth [MBytes/s]' in mem_results: self.results['MEM BW [MByte/s]'] = mem_results['Memory bandwidth [MBytes/s]'] else: self.results['MEM BW [MByte/s]'] = mem_results['Memory BW [MBytes/s]'] self.results['Performance [MLUP/s]'] = ( iterations_per_repetition / time_per_repetition) / 1e6 self.results['Performance [MIt/s]'] = ( iterations_per_repetition / time_per_repetition) / 1e6
Report gathered analysis data in human readable form.
def report(self, output_file=sys.stdout): """Report gathered analysis data in human readable form.""" if self.verbose > 1: with pprint_nosort(): pprint.pprint(self.results) if self.verbose > 0: print('Runtime (per repetition): {:.2g} s'.format( self.results['Runtime (per repetition) [s]']), file=output_file) if self.verbose > 0: print('Iterations per repetition: {!s}'.format( self.results['Iterations per repetition']), file=output_file) print('Runtime (per cacheline update): {:.2f} cy/CL'.format( self.results['Runtime (per cacheline update) [cy/CL]']), file=output_file) print('MEM volume (per repetition): {:.0f} Byte'.format( self.results['MEM volume (per repetition) [B]']), file=output_file) print('Performance: {:.2f} MFLOP/s'.format(self.results['Performance [MFLOP/s]']), file=output_file) print('Performance: {:.2f} MLUP/s'.format(self.results['Performance [MLUP/s]']), file=output_file) print('Performance: {:.2f} It/s'.format(self.results['Performance [MIt/s]']), file=output_file) if self.verbose > 0: print('MEM bandwidth: {:.2f} MByte/s'.format(self.results['MEM BW [MByte/s]']), file=output_file) print('', file=output_file) if not self.no_phenoecm: print("Data Transfers:") print("{:^8} |".format("cache"), end='') for metrics in self.results['data transfers'].values(): for metric_name in sorted(metrics): print(" {:^14}".format(metric_name), end='') print() break for cache, metrics in sorted(self.results['data transfers'].items()): print("{!s:^8} |".format(cache), end='') for k, v in sorted(metrics.items()): print(" {!s:^14}".format(v), end='') print() print() print('Phenomenological ECM model: {{ {T_OL:.1f} || {T_nOL:.1f} | {T_L1L2:.1f} | ' '{T_L2L3:.1f} | {T_L3MEM:.1f} }} cy/CL'.format( **{k: float(v) for k, v in self.results['ECM'].items()}), file=output_file) print('T_OL assumes that two loads per cycle may be retiered, which is true for ' '128bit SSE/half-AVX loads on SNB and IVY, and 256bit full-AVX loads on HSW, ' 'BDW, SKL and SKX, but it also depends on AGU availability.', file=output_file)
Parse the description in the README file
def parse_description(): """ Parse the description in the README file CommandLine: python -c "import setup; print(setup.parse_description())" """ from os.path import dirname, join, exists readme_fpath = join(dirname(__file__), 'README.md') # print('readme_fpath = %r' % (readme_fpath,)) # This breaks on pip install, so check that it exists. if exists(readme_fpath): # try: # # convert markdown to rst for pypi # import pypandoc # return pypandoc.convert(readme_fpath, 'rst') # except Exception as ex: # strip out markdown to make a clean readme for pypi textlines = [] with open(readme_fpath, 'r') as f: capture = False for line in f.readlines(): if '# Purpose' in line: capture = True elif line.startswith('##'): break elif capture: textlines += [line] text = ''.join(textlines).strip() text = text.replace('\n\n', '_NLHACK_') text = text.replace('\n', ' ') text = text.replace('_NLHACK_', '\n\n') return text return ''
Schedule a retry
def schedule_retry(self, config): """Schedule a retry""" raise self.retry(countdown=config.get('SAILTHRU_RETRY_SECONDS'), max_retries=config.get('SAILTHRU_RETRY_ATTEMPTS'))
Build and return Sailthru purchase item object
def _build_purchase_item(course_id, course_url, cost_in_cents, mode, course_data, sku): """Build and return Sailthru purchase item object""" # build item description item = { 'id': "{}-{}".format(course_id, mode), 'url': course_url, 'price': cost_in_cents, 'qty': 1, } # get title from course info if we don't already have it from Sailthru if 'title' in course_data: item['title'] = course_data['title'] else: # can't find, just invent title item['title'] = 'Course {} mode: {}'.format(course_id, mode) if 'tags' in course_data: item['tags'] = course_data['tags'] # add vars to item item['vars'] = dict(course_data.get('vars', {}), mode=mode, course_run_id=course_id) item['vars']['purchase_sku'] = sku return item
Record a purchase in Sailthru
def _record_purchase(sailthru_client, email, item, purchase_incomplete, message_id, options): """Record a purchase in Sailthru Arguments: sailthru_client (object): SailthruClient email (str): user's email address item (dict): Sailthru required information about the course purchase_incomplete (boolean): True if adding item to shopping cart message_id (str): Cookie used to identify marketing campaign options (dict): Sailthru purchase API options (e.g. template name) Returns: False if retryable error, else True """ try: sailthru_response = sailthru_client.purchase(email, [item], incomplete=purchase_incomplete, message_id=message_id, options=options) if not sailthru_response.is_ok(): error = sailthru_response.get_error() logger.error("Error attempting to record purchase in Sailthru: %s", error.get_message()) return not can_retry_sailthru_request(error) except SailthruClientError as exc: logger.exception("Exception attempting to record purchase for %s in Sailthru - %s", email, text_type(exc)) return False return True
Get course information using the Sailthru content api or from cache.
def _get_course_content(course_id, course_url, sailthru_client, site_code, config): """Get course information using the Sailthru content api or from cache. If there is an error, just return with an empty response. Arguments: course_id (str): course key of the course course_url (str): LMS url for course info page. sailthru_client (object): SailthruClient site_code (str): site code config (dict): config options Returns: course information from Sailthru """ # check cache first cache_key = "{}:{}".format(site_code, course_url) response = cache.get(cache_key) if not response: try: sailthru_response = sailthru_client.api_get("content", {"id": course_url}) if not sailthru_response.is_ok(): response = {} else: response = sailthru_response.json cache.set(cache_key, response, config.get('SAILTHRU_CACHE_TTL_SECONDS')) except SailthruClientError: response = {} if not response: logger.error('Could not get course data from Sailthru on enroll/purchase event. ' 'Calling Ecommerce Course API to get course info for enrollment confirmation email') response = _get_course_content_from_ecommerce(course_id, site_code=site_code) if response: cache.set(cache_key, response, config.get('SAILTHRU_CACHE_TTL_SECONDS')) return response
Get course information using the Ecommerce course api.
def _get_course_content_from_ecommerce(course_id, site_code=None): """ Get course information using the Ecommerce course api. In case of error returns empty response. Arguments: course_id (str): course key of the course site_code (str): site code Returns: course information from Ecommerce """ api = get_ecommerce_client(site_code=site_code) try: api_response = api.courses(course_id).get() except Exception: # pylint: disable=broad-except logger.exception( 'An error occurred while retrieving data for course run [%s] from the Catalog API.', course_id, exc_info=True ) return {} return { 'title': api_response.get('name'), 'verification_deadline': api_response.get('verification_deadline') }
Maintain a list of courses the user has unenrolled from in the Sailthru user record
def _update_unenrolled_list(sailthru_client, email, course_url, unenroll): """Maintain a list of courses the user has unenrolled from in the Sailthru user record Arguments: sailthru_client (object): SailthruClient email (str): user's email address course_url (str): LMS url for course info page. unenroll (boolean): True if unenrolling, False if enrolling Returns: False if retryable error, else True """ try: # get the user 'vars' values from sailthru sailthru_response = sailthru_client.api_get("user", {"id": email, "fields": {"vars": 1}}) if not sailthru_response.is_ok(): error = sailthru_response.get_error() logger.error("Error attempting to read user record from Sailthru: %s", error.get_message()) return not can_retry_sailthru_request(error) response_json = sailthru_response.json unenroll_list = [] if response_json and "vars" in response_json and response_json["vars"] \ and "unenrolled" in response_json["vars"]: unenroll_list = response_json["vars"]["unenrolled"] changed = False # if unenrolling, add course to unenroll list if unenroll: if course_url not in unenroll_list: unenroll_list.append(course_url) changed = True # if enrolling, remove course from unenroll list elif course_url in unenroll_list: unenroll_list.remove(course_url) changed = True if changed: # write user record back sailthru_response = sailthru_client.api_post( 'user', {'id': email, 'key': 'email', 'vars': {'unenrolled': unenroll_list}}) if not sailthru_response.is_ok(): error = sailthru_response.get_error() logger.error("Error attempting to update user record in Sailthru: %s", error.get_message()) return not can_retry_sailthru_request(error) return True except SailthruClientError as exc: logger.exception("Exception attempting to update user record for %s in Sailthru - %s", email, text_type(exc)) return False
Adds/ updates Sailthru when a user adds to cart/ purchases/ upgrades a course
def update_course_enrollment(self, email, course_url, purchase_incomplete, mode, unit_cost=None, course_id=None, currency=None, message_id=None, site_code=None, sku=None): """Adds/updates Sailthru when a user adds to cart/purchases/upgrades a course Args: email(str): The user's email address course_url(str): Course home page url purchase_incomplete(boolean): True if adding to cart mode(string): enroll mode (audit, verification, ...) unit_cost(decimal): cost if purchase event course_id(CourseKey): course id currency(str): currency if purchase event - currently ignored since Sailthru only supports USD message_id(str): value from Sailthru marketing campaign cookie site_code(str): site code Returns: None """ # Get configuration config = get_sailthru_configuration(site_code) try: sailthru_client = get_sailthru_client(site_code) except SailthruError: # NOTE: We rely on the function to log the error for us return # Use event type to figure out processing required new_enroll = False send_template = None if not purchase_incomplete: if mode == 'verified': # upgrade complete send_template = config.get('SAILTHRU_UPGRADE_TEMPLATE') elif mode == 'audit' or mode == 'honor': # free enroll new_enroll = True send_template = config.get('SAILTHRU_ENROLL_TEMPLATE') else: # paid course purchase complete new_enroll = True send_template = config.get('SAILTHRU_PURCHASE_TEMPLATE') # calc price in pennies for Sailthru # https://getstarted.sailthru.com/new-for-developers-overview/advanced-features/purchase/ cost_in_cents = int(unit_cost * 100) # update the "unenrolled" course array in the user record on Sailthru if new enroll or unenroll if new_enroll: if not _update_unenrolled_list(sailthru_client, email, course_url, False): schedule_retry(self, config) # Get course data from Sailthru content library or cache course_data = _get_course_content(course_id, course_url, sailthru_client, site_code, config) # build item description item = _build_purchase_item(course_id, course_url, cost_in_cents, mode, course_data, sku) # build purchase api options list options = {} if purchase_incomplete and config.get('SAILTHRU_ABANDONED_CART_TEMPLATE'): options['reminder_template'] = config.get('SAILTHRU_ABANDONED_CART_TEMPLATE') # Sailthru reminder time format is '+n time unit' options['reminder_time'] = "+{} minutes".format(config.get('SAILTHRU_ABANDONED_CART_DELAY')) # add appropriate send template if send_template: options['send_template'] = send_template if not _record_purchase(sailthru_client, email, item, purchase_incomplete, message_id, options): schedule_retry(self, config)
Sends the course refund email.
def send_course_refund_email(self, email, refund_id, amount, course_name, order_number, order_url, site_code=None): """ Sends the course refund email. Args: self: Ignore. email (str): Recipient's email address. refund_id (int): ID of the refund that initiated this task. amount (str): Formatted amount of the refund. course_name (str): Name of the course for which payment was refunded. order_number (str): Order number of the order that was refunded. order_url (str): Receipt URL of the refunded order. site_code (str): Identifier of the site sending the email. """ config = get_sailthru_configuration(site_code) try: sailthru_client = get_sailthru_client(site_code) except SailthruError: # NOTE: We rely on the function to log the error for us return email_vars = { 'amount': amount, 'course_name': course_name, 'order_number': order_number, 'order_url': order_url, } try: response = sailthru_client.send( template=config['templates']['course_refund'], email=email, _vars=email_vars ) except SailthruClientError: logger.exception( 'A client error occurred while attempting to send a course refund notification for refund [%d].', refund_id ) return if response.is_ok(): logger.info('Course refund notification sent for refund %d.', refund_id) else: error = response.get_error() logger.error( 'An error occurred while attempting to send a course refund notification for refund [%d]: %d - %s', refund_id, error.get_error_code(), error.get_message() ) if can_retry_sailthru_request(error): logger.info( 'An attempt will be made again to send a course refund notification for refund [%d].', refund_id ) schedule_retry(self, config) else: logger.warning( 'No further attempts will be made to send a course refund notification for refund [%d].', refund_id )
Sends the offer assignment email. Args: self: Ignore. user_email ( str ): Recipient s email address. offer_assignment_id ( str ): Key of the entry in the offer_assignment model. subject ( str ): Email subject. email_body ( str ): The body of the email. site_code ( str ): Identifier of the site sending the email.
def send_offer_assignment_email(self, user_email, offer_assignment_id, subject, email_body, site_code=None): """ Sends the offer assignment email. Args: self: Ignore. user_email (str): Recipient's email address. offer_assignment_id (str): Key of the entry in the offer_assignment model. subject (str): Email subject. email_body (str): The body of the email. site_code (str): Identifier of the site sending the email. """ config = get_sailthru_configuration(site_code) response = _send_offer_assignment_notification_email(config, user_email, subject, email_body, site_code, self) if response and response.is_ok(): send_id = response.get_body().get('send_id') # pylint: disable=no-member if _update_assignment_email_status(offer_assignment_id, send_id, 'success'): logger.info('[Offer Assignment] Offer assignment notification sent with message --- {message}'.format( message=email_body)) else: logger.exception( '[Offer Assignment] An error occurred while updating email status data for ' 'offer {token_offer} and email {token_email} via the ecommerce API.'.format( token_offer=offer_assignment_id, token_email=user_email, ) )
Handles sending offer assignment notification emails and retrying failed emails when appropriate.
def _send_offer_assignment_notification_email(config, user_email, subject, email_body, site_code, task): """Handles sending offer assignment notification emails and retrying failed emails when appropriate.""" try: sailthru_client = get_sailthru_client(site_code) except SailthruError: logger.exception( '[Offer Assignment] A client error occurred while attempting to send a offer assignment notification.' ' Message: {message}'.format(message=email_body) ) return None email_vars = { 'subject': subject, 'email_body': email_body, } try: response = sailthru_client.send( template=config['templates']['assignment_email'], email=user_email, _vars=email_vars ) except SailthruClientError: logger.exception( '[Offer Assignment] A client error occurred while attempting to send a offer assignment notification.' ' Message: {message}'.format(message=email_body) ) return None if not response.is_ok(): error = response.get_error() logger.error( '[Offer Assignment] A {token_error_code} - {token_error_message} error occurred' ' while attempting to send a offer assignment notification.' ' Message: {message}'.format( message=email_body, token_error_code=error.get_error_code(), token_error_message=error.get_message() ) ) if can_retry_sailthru_request(error): logger.info( '[Offer Assignment] An attempt will be made to resend the offer assignment notification.' ' Message: {message}'.format(message=email_body) ) schedule_retry(task, config) else: logger.warning( '[Offer Assignment] No further attempts will be made to send the offer assignment notification.' ' Failed Message: {message}'.format(message=email_body) ) return response
Update the offer_assignment and offer_assignment_email model using the Ecommerce assignmentemail api. Arguments: offer_assignment_id ( str ): Key of the entry in the offer_assignment model. send_id ( str ): Unique message id from Sailthru status ( str ): status to be sent to the api site_code ( str ): site code Returns: True or False based on model update status from Ecommerce api
def _update_assignment_email_status(offer_assignment_id, send_id, status, site_code=None): """ Update the offer_assignment and offer_assignment_email model using the Ecommerce assignmentemail api. Arguments: offer_assignment_id (str): Key of the entry in the offer_assignment model. send_id (str): Unique message id from Sailthru status (str): status to be sent to the api site_code (str): site code Returns: True or False based on model update status from Ecommerce api """ api = get_ecommerce_client(url_postfix='assignment-email/', site_code=site_code) post_data = { 'offer_assignment_id': offer_assignment_id, 'send_id': send_id, 'status': status, } try: api_response = api.status().post(post_data) except RequestException: logger.exception( '[Offer Assignment] An error occurred while updating offer assignment email status for ' 'offer id {token_offer} and message id {token_send_id} via the Ecommerce API.'.format( token_offer=offer_assignment_id, token_send_id=send_id ) ) return False return True if api_response.get('status') == 'updated' else False
Sends the offer emails after assignment either for revoking or reminding. Args: self: Ignore. user_email ( str ): Recipient s email address. subject ( str ): Email subject. email_body ( str ): The body of the email. site_code ( str ): Identifier of the site sending the email.
def send_offer_update_email(self, user_email, subject, email_body, site_code=None): """ Sends the offer emails after assignment, either for revoking or reminding. Args: self: Ignore. user_email (str): Recipient's email address. subject (str): Email subject. email_body (str): The body of the email. site_code (str): Identifier of the site sending the email. """ config = get_sailthru_configuration(site_code) _send_offer_assignment_notification_email(config, user_email, subject, email_body, site_code, self)
Returns a dictionary containing logging configuration.
def get_logger_config(log_dir='/var/tmp', logging_env='no_env', edx_filename='edx.log', dev_env=False, debug=False, local_loglevel='INFO', service_variant='ecomworker'): """ Returns a dictionary containing logging configuration. If dev_env is True, logging will not be done via local rsyslogd. Instead, application logs will be dropped into log_dir. 'edx_filename' is ignored unless dev_env is True. """ # Revert to INFO if an invalid string is passed in if local_loglevel not in ['DEBUG', 'INFO', 'WARNING', 'ERROR', 'CRITICAL']: local_loglevel = 'INFO' hostname = platform.node().split('.')[0] syslog_format = ( '[service_variant={service_variant}]' '[%(name)s][env:{logging_env}] %(levelname)s ' '[{hostname} %(process)d] [%(filename)s:%(lineno)d] ' '- %(message)s' ).format( service_variant=service_variant, logging_env=logging_env, hostname=hostname ) if debug: handlers = ['console'] else: handlers = ['local'] logger_config = { 'version': 1, 'disable_existing_loggers': False, 'formatters': { 'standard': { 'format': '%(asctime)s %(levelname)s %(process)d ' '[%(name)s] %(filename)s:%(lineno)d - %(message)s', }, 'syslog_format': {'format': syslog_format}, 'raw': {'format': '%(message)s'}, }, 'handlers': { 'console': { 'level': 'DEBUG' if debug else 'INFO', 'class': 'logging.StreamHandler', 'formatter': 'standard', 'stream': sys.stdout, }, }, 'loggers': { 'requests': { 'handlers': handlers, 'level': 'WARNING', 'propagate': True }, '': { 'handlers': handlers, 'level': 'DEBUG', 'propagate': False }, } } if dev_env: edx_file_loc = os.path.join(log_dir, edx_filename) logger_config['handlers'].update({ 'local': { 'class': 'logging.handlers.RotatingFileHandler', 'level': local_loglevel, 'formatter': 'standard', 'filename': edx_file_loc, 'maxBytes': 1024 * 1024 * 2, 'backupCount': 5, }, }) else: logger_config['handlers'].update({ 'local': { 'level': local_loglevel, 'class': 'logging.handlers.SysLogHandler', # Use a different address for Mac OS X 'address': '/var/run/syslog' if sys.platform == 'darwin' else '/dev/log', 'formatter': 'syslog_format', 'facility': SysLogHandler.LOG_LOCAL0, }, }) return logger_config
Retry with exponential backoff until fulfillment succeeds or the retry limit is reached. If the retry limit is exceeded the exception is re - raised.
def _retry_order(self, exception, max_fulfillment_retries, order_number): """ Retry with exponential backoff until fulfillment succeeds or the retry limit is reached. If the retry limit is exceeded, the exception is re-raised. """ retries = self.request.retries if retries == max_fulfillment_retries: logger.exception('Fulfillment of order [%s] failed. Giving up.', order_number) else: logger.warning('Fulfillment of order [%s] failed. Retrying.', order_number) countdown = 2 ** retries raise self.retry(exc=exception, countdown=countdown, max_retries=max_fulfillment_retries)
Fulfills an order.
def fulfill_order(self, order_number, site_code=None, email_opt_in=False): """Fulfills an order. Arguments: order_number (str): Order number indicating which order to fulfill. Returns: None """ max_fulfillment_retries = get_configuration('MAX_FULFILLMENT_RETRIES', site_code=site_code) api = get_ecommerce_client(site_code=site_code) try: logger.info('Requesting fulfillment of order [%s].', order_number) api.orders(order_number).fulfill.put(email_opt_in=email_opt_in) except exceptions.HttpClientError as exc: status_code = exc.response.status_code # pylint: disable=no-member if status_code == 406: # The order is not fulfillable. Therefore, it must be complete. logger.info('Order [%s] has already been fulfilled. Ignoring.', order_number) raise Ignore() else: # Unknown client error. Let's retry to resolve it. logger.warning( 'Fulfillment of order [%s] failed because of HttpClientError. Retrying', order_number, exc_info=True ) _retry_order(self, exc, max_fulfillment_retries, order_number) except (exceptions.HttpServerError, exceptions.Timeout, SSLError) as exc: # Fulfillment failed, retry _retry_order(self, exc, max_fulfillment_retries, order_number)
Returns a Sailthru client for the specified site.
def get_sailthru_client(site_code): """ Returns a Sailthru client for the specified site. Args: site_code (str): Site for which the client should be configured. Returns: SailthruClient Raises: SailthruNotEnabled: If Sailthru is not enabled for the specified site. ConfigurationError: If either the Sailthru API key or secret are not set for the site. """ # Get configuration config = get_sailthru_configuration(site_code) # Return if Sailthru integration disabled if not config.get('SAILTHRU_ENABLE'): msg = 'Sailthru is not enabled for site {}'.format(site_code) log.debug(msg) raise SailthruNotEnabled(msg) # Make sure key and secret configured key = config.get('SAILTHRU_KEY') secret = config.get('SAILTHRU_SECRET') if not (key and secret): msg = 'Both key and secret are required for site {}'.format(site_code) log.error(msg) raise ConfigurationError(msg) return SailthruClient(key, secret)
Get an object from the cache
def get(self, key): """Get an object from the cache Arguments: key (str): Cache key Returns: Cached object """ lock.acquire() try: if key not in self: return None current_time = time.time() if self[key].expire > current_time: return self[key].value # expired key, clean out all expired keys deletes = [] for k, val in self.items(): if val.expire <= current_time: deletes.append(k) for k in deletes: del self[k] return None finally: lock.release()
Save an object in the cache
def set(self, key, value, duration): """Save an object in the cache Arguments: key (str): Cache key value (object): object to cache duration (int): time in seconds to keep object in cache """ lock.acquire() try: self[key] = CacheObject(value, duration) finally: lock.release()
Get a value from configuration.
def get_configuration(variable, site_code=None): """ Get a value from configuration. Retrieves the value corresponding to the given variable from the configuration module currently in use by the app. Specify a site_code value to check for a site-specific override. Arguments: variable (str): The name of a variable from the configuration module. Keyword Arguments: site_code (str): The SITE_OVERRIDES key to inspect for site-specific values Returns: The value corresponding to the variable, or None if the variable is not found. """ name = os.environ.get(CONFIGURATION_MODULE) # __import__ performs a full import, but only returns the top-level # package, not the targeted module. sys.modules is a dictionary # mapping module names to loaded modules. __import__(name) module = sys.modules[name] # Locate the setting in the specified module, then attempt to apply a site-specific override setting_value = getattr(module, variable, None) site_overrides = getattr(module, 'SITE_OVERRIDES', None) if site_overrides and site_code is not None: site_specific_overrides = site_overrides.get(site_code) if site_specific_overrides: override_value = site_specific_overrides.get(variable) if override_value: setting_value = override_value if setting_value is None: raise RuntimeError('Worker is improperly configured: {} is unset in {}.'.format(variable, module)) return setting_value
Get client for fetching data from ecommerce API. Arguments: site_code ( str ): ( Optional ) The SITE_OVERRIDES key to inspect for site - specific values url_postfix ( str ): ( Optional ) The URL postfix value to append to the ECOMMERCE_API_ROOT value.
def get_ecommerce_client(url_postfix='', site_code=None): """ Get client for fetching data from ecommerce API. Arguments: site_code (str): (Optional) The SITE_OVERRIDES key to inspect for site-specific values url_postfix (str): (Optional) The URL postfix value to append to the ECOMMERCE_API_ROOT value. Returns: EdxRestApiClient object """ ecommerce_api_root = get_configuration('ECOMMERCE_API_ROOT', site_code=site_code) signing_key = get_configuration('JWT_SECRET_KEY', site_code=site_code) issuer = get_configuration('JWT_ISSUER', site_code=site_code) service_username = get_configuration('ECOMMERCE_SERVICE_USERNAME', site_code=site_code) return EdxRestApiClient( ecommerce_api_root + url_postfix, signing_key=signing_key, issuer=issuer, username=service_username)