repository_name
stringlengths
7
55
func_path_in_repository
stringlengths
4
223
func_name
stringlengths
1
134
whole_func_string
stringlengths
75
104k
language
stringclasses
1 value
func_code_string
stringlengths
75
104k
func_code_tokens
listlengths
19
28.4k
func_documentation_string
stringlengths
1
46.9k
func_documentation_tokens
listlengths
1
1.97k
split_name
stringclasses
1 value
func_code_url
stringlengths
87
315
RaRe-Technologies/gensim-simserver
simserver/simserver.py
SessionServer.terminate
def terminate(self): """Delete all files created by this server, invalidating `self`. Use with care.""" logger.info("deleting entire server %s" % self) self.close() try: shutil.rmtree(self.basedir) logger.info("deleted server under %s" % self.basedir) # delete everything from self, so that using this object fails results # in an error as quickly as possible for val in self.__dict__.keys(): try: delattr(self, val) except: pass except Exception, e: logger.warning("failed to delete SessionServer: %s" % (e))
python
def terminate(self): """Delete all files created by this server, invalidating `self`. Use with care.""" logger.info("deleting entire server %s" % self) self.close() try: shutil.rmtree(self.basedir) logger.info("deleted server under %s" % self.basedir) # delete everything from self, so that using this object fails results # in an error as quickly as possible for val in self.__dict__.keys(): try: delattr(self, val) except: pass except Exception, e: logger.warning("failed to delete SessionServer: %s" % (e))
[ "def", "terminate", "(", "self", ")", ":", "logger", ".", "info", "(", "\"deleting entire server %s\"", "%", "self", ")", "self", ".", "close", "(", ")", "try", ":", "shutil", ".", "rmtree", "(", "self", ".", "basedir", ")", "logger", ".", "info", "(", "\"deleted server under %s\"", "%", "self", ".", "basedir", ")", "# delete everything from self, so that using this object fails results", "# in an error as quickly as possible", "for", "val", "in", "self", ".", "__dict__", ".", "keys", "(", ")", ":", "try", ":", "delattr", "(", "self", ",", "val", ")", "except", ":", "pass", "except", "Exception", ",", "e", ":", "logger", ".", "warning", "(", "\"failed to delete SessionServer: %s\"", "%", "(", "e", ")", ")" ]
Delete all files created by this server, invalidating `self`. Use with care.
[ "Delete", "all", "files", "created", "by", "this", "server", "invalidating", "self", ".", "Use", "with", "care", "." ]
train
https://github.com/RaRe-Technologies/gensim-simserver/blob/e7e59e836ef6d9da019a8c6b218ef0bdd998b2da/simserver/simserver.py#L1024-L1039
RaRe-Technologies/gensim-simserver
simserver/simserver.py
SessionServer.find_similar
def find_similar(self, *args, **kwargs): """ Find similar articles. With autosession off, use the index state *before* current session started, so that changes made in the session will not be visible here. With autosession on, close the current session first (so that session changes *are* committed and visible). """ if self.session is not None and self.autosession: # with autosession on, commit the pending transaction first self.commit() return self.stable.find_similar(*args, **kwargs)
python
def find_similar(self, *args, **kwargs): """ Find similar articles. With autosession off, use the index state *before* current session started, so that changes made in the session will not be visible here. With autosession on, close the current session first (so that session changes *are* committed and visible). """ if self.session is not None and self.autosession: # with autosession on, commit the pending transaction first self.commit() return self.stable.find_similar(*args, **kwargs)
[ "def", "find_similar", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "self", ".", "session", "is", "not", "None", "and", "self", ".", "autosession", ":", "# with autosession on, commit the pending transaction first", "self", ".", "commit", "(", ")", "return", "self", ".", "stable", ".", "find_similar", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
Find similar articles. With autosession off, use the index state *before* current session started, so that changes made in the session will not be visible here. With autosession on, close the current session first (so that session changes *are* committed and visible).
[ "Find", "similar", "articles", "." ]
train
https://github.com/RaRe-Technologies/gensim-simserver/blob/e7e59e836ef6d9da019a8c6b218ef0bdd998b2da/simserver/simserver.py#L1042-L1054
cree-py/pynite
examples/discord.py
Fortnite.profile
async def profile(self, ctx, platform, name): '''Fetch a profile.''' player = await self.client.get_player(platform, name) solos = await player.get_solos() await ctx.send("# of kills in solos for {}: {}".format(name,solos.kills.value))
python
async def profile(self, ctx, platform, name): '''Fetch a profile.''' player = await self.client.get_player(platform, name) solos = await player.get_solos() await ctx.send("# of kills in solos for {}: {}".format(name,solos.kills.value))
[ "async", "def", "profile", "(", "self", ",", "ctx", ",", "platform", ",", "name", ")", ":", "player", "=", "await", "self", ".", "client", ".", "get_player", "(", "platform", ",", "name", ")", "solos", "=", "await", "player", ".", "get_solos", "(", ")", "await", "ctx", ".", "send", "(", "\"# of kills in solos for {}: {}\"", ".", "format", "(", "name", ",", "solos", ".", "kills", ".", "value", ")", ")" ]
Fetch a profile.
[ "Fetch", "a", "profile", "." ]
train
https://github.com/cree-py/pynite/blob/91424c912a23d2b93bda1f017703e492572134c6/examples/discord.py#L22-L28
anti1869/aiohttp_autoreload
src/aiohttp_autoreload.py
start
def start(io_loop=None, check_time=2): """Begins watching source files for changes. .. versionchanged:: 4.1 The ``io_loop`` argument is deprecated. """ io_loop = io_loop or asyncio.get_event_loop() if io_loop in _io_loops: return _io_loops[io_loop] = True if len(_io_loops) > 1: logger.warning("aiohttp_autoreload started more than once in the same process") # if _has_execv: # add_reload_hook(functools.partial(io_loop.close, all_fds=True)) modify_times = {} callback = functools.partial(_reload_on_update, modify_times) logger.debug("Starting periodic checks for code changes") call_periodic(check_time, callback, loop=io_loop)
python
def start(io_loop=None, check_time=2): """Begins watching source files for changes. .. versionchanged:: 4.1 The ``io_loop`` argument is deprecated. """ io_loop = io_loop or asyncio.get_event_loop() if io_loop in _io_loops: return _io_loops[io_loop] = True if len(_io_loops) > 1: logger.warning("aiohttp_autoreload started more than once in the same process") # if _has_execv: # add_reload_hook(functools.partial(io_loop.close, all_fds=True)) modify_times = {} callback = functools.partial(_reload_on_update, modify_times) logger.debug("Starting periodic checks for code changes") call_periodic(check_time, callback, loop=io_loop)
[ "def", "start", "(", "io_loop", "=", "None", ",", "check_time", "=", "2", ")", ":", "io_loop", "=", "io_loop", "or", "asyncio", ".", "get_event_loop", "(", ")", "if", "io_loop", "in", "_io_loops", ":", "return", "_io_loops", "[", "io_loop", "]", "=", "True", "if", "len", "(", "_io_loops", ")", ">", "1", ":", "logger", ".", "warning", "(", "\"aiohttp_autoreload started more than once in the same process\"", ")", "# if _has_execv:", "# add_reload_hook(functools.partial(io_loop.close, all_fds=True))", "modify_times", "=", "{", "}", "callback", "=", "functools", ".", "partial", "(", "_reload_on_update", ",", "modify_times", ")", "logger", ".", "debug", "(", "\"Starting periodic checks for code changes\"", ")", "call_periodic", "(", "check_time", ",", "callback", ",", "loop", "=", "io_loop", ")" ]
Begins watching source files for changes. .. versionchanged:: 4.1 The ``io_loop`` argument is deprecated.
[ "Begins", "watching", "source", "files", "for", "changes", "." ]
train
https://github.com/anti1869/aiohttp_autoreload/blob/1d9b9b67ffba1073fa35495d87538b4570807367/src/aiohttp_autoreload.py#L58-L76
google/dotty
efilter/protocols/reducer.py
generate_chunks
def generate_chunks(data, chunk_size=DEFAULT_CHUNK_SIZE): """Yield 'chunk_size' items from 'data' at a time.""" iterator = iter(repeated.getvalues(data)) while True: chunk = list(itertools.islice(iterator, chunk_size)) if not chunk: return yield chunk
python
def generate_chunks(data, chunk_size=DEFAULT_CHUNK_SIZE): """Yield 'chunk_size' items from 'data' at a time.""" iterator = iter(repeated.getvalues(data)) while True: chunk = list(itertools.islice(iterator, chunk_size)) if not chunk: return yield chunk
[ "def", "generate_chunks", "(", "data", ",", "chunk_size", "=", "DEFAULT_CHUNK_SIZE", ")", ":", "iterator", "=", "iter", "(", "repeated", ".", "getvalues", "(", "data", ")", ")", "while", "True", ":", "chunk", "=", "list", "(", "itertools", ".", "islice", "(", "iterator", ",", "chunk_size", ")", ")", "if", "not", "chunk", ":", "return", "yield", "chunk" ]
Yield 'chunk_size' items from 'data' at a time.
[ "Yield", "chunk_size", "items", "from", "data", "at", "a", "time", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/protocols/reducer.py#L58-L67
google/dotty
efilter/protocols/reducer.py
reduce
def reduce(reducer, data, chunk_size=DEFAULT_CHUNK_SIZE): """Repeatedly call fold and merge on data and then finalize. Arguments: data: Input for the fold function. reducer: The IReducer to use. chunk_size: How many items should be passed to fold at a time? Returns: Return value of finalize. """ if not chunk_size: return finalize(reducer, fold(reducer, data)) # Splitting the work up into chunks allows us to, e.g. reduce a large file # without loading everything into memory, while still being significantly # faster than repeatedly calling the fold function for every element. chunks = generate_chunks(data, chunk_size) intermediate = fold(reducer, next(chunks)) for chunk in chunks: intermediate = merge(reducer, intermediate, fold(reducer, chunk)) return finalize(reducer, intermediate)
python
def reduce(reducer, data, chunk_size=DEFAULT_CHUNK_SIZE): """Repeatedly call fold and merge on data and then finalize. Arguments: data: Input for the fold function. reducer: The IReducer to use. chunk_size: How many items should be passed to fold at a time? Returns: Return value of finalize. """ if not chunk_size: return finalize(reducer, fold(reducer, data)) # Splitting the work up into chunks allows us to, e.g. reduce a large file # without loading everything into memory, while still being significantly # faster than repeatedly calling the fold function for every element. chunks = generate_chunks(data, chunk_size) intermediate = fold(reducer, next(chunks)) for chunk in chunks: intermediate = merge(reducer, intermediate, fold(reducer, chunk)) return finalize(reducer, intermediate)
[ "def", "reduce", "(", "reducer", ",", "data", ",", "chunk_size", "=", "DEFAULT_CHUNK_SIZE", ")", ":", "if", "not", "chunk_size", ":", "return", "finalize", "(", "reducer", ",", "fold", "(", "reducer", ",", "data", ")", ")", "# Splitting the work up into chunks allows us to, e.g. reduce a large file", "# without loading everything into memory, while still being significantly", "# faster than repeatedly calling the fold function for every element.", "chunks", "=", "generate_chunks", "(", "data", ",", "chunk_size", ")", "intermediate", "=", "fold", "(", "reducer", ",", "next", "(", "chunks", ")", ")", "for", "chunk", "in", "chunks", ":", "intermediate", "=", "merge", "(", "reducer", ",", "intermediate", ",", "fold", "(", "reducer", ",", "chunk", ")", ")", "return", "finalize", "(", "reducer", ",", "intermediate", ")" ]
Repeatedly call fold and merge on data and then finalize. Arguments: data: Input for the fold function. reducer: The IReducer to use. chunk_size: How many items should be passed to fold at a time? Returns: Return value of finalize.
[ "Repeatedly", "call", "fold", "and", "merge", "on", "data", "and", "then", "finalize", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/protocols/reducer.py#L70-L92
google/dotty
efilter/ast.py
IfElse.conditions
def conditions(self): """The if-else pairs.""" for idx in six.moves.range(1, len(self.children), 2): yield (self.children[idx - 1], self.children[idx])
python
def conditions(self): """The if-else pairs.""" for idx in six.moves.range(1, len(self.children), 2): yield (self.children[idx - 1], self.children[idx])
[ "def", "conditions", "(", "self", ")", ":", "for", "idx", "in", "six", ".", "moves", ".", "range", "(", "1", ",", "len", "(", "self", ".", "children", ")", ",", "2", ")", ":", "yield", "(", "self", ".", "children", "[", "idx", "-", "1", "]", ",", "self", ".", "children", "[", "idx", "]", ")" ]
The if-else pairs.
[ "The", "if", "-", "else", "pairs", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/ast.py#L415-L418
radical-cybertools/radical.entk
src/radical/entk/execman/rp/task_processor.py
resolve_placeholders
def resolve_placeholders(path, placeholder_dict): """ **Purpose**: Substitute placeholders in staging attributes of a Task with actual paths to the corresponding tasks. :arguments: :path: string describing the staging paths, possibly containing a placeholder :placeholder_dict: dictionary holding the values for placeholders """ try: if isinstance(path, unicode): path = str(path) if not isinstance(path, str): raise TypeError(expected_type=str, actual_type=type(path)) if '$' not in path: return path # Extract placeholder from path if len(path.split('>')) == 1: placeholder = path.split('/')[0] else: if path.split('>')[0].strip().startswith('$'): placeholder = path.split('>')[0].strip().split('/')[0] else: placeholder = path.split('>')[1].strip().split('/')[0] # SHARED if placeholder == "$SHARED": return path.replace(placeholder, 'pilot://') # Expected placeholder format: # $Pipeline_{pipeline.uid}_Stage_{stage.uid}_Task_{task.uid} broken_placeholder = placeholder.split('/')[0].split('_') if not len(broken_placeholder) == 6: raise ValueError( obj='placeholder', attribute='task', expected_value='$Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name) or $SHARED', actual_value=broken_placeholder) pipeline_name = broken_placeholder[1] stage_name = broken_placeholder[3] task_name = broken_placeholder[5] resolved_placeholder = None if pipeline_name in placeholder_dict.keys(): if stage_name in placeholder_dict[pipeline_name].keys(): if task_name in placeholder_dict[pipeline_name][stage_name].keys(): resolved_placeholder = path.replace(placeholder, placeholder_dict[ pipeline_name][stage_name][task_name]['path']) else: logger.warning('%s not assigned to any task in Stage %s Pipeline %s' % (task_name, stage_name, pipeline_name)) else: logger.warning('%s not assigned to any Stage in Pipeline %s' % ( stage_name, pipeline_name)) else: logger.warning('%s not assigned to any Pipeline' % (pipeline_name)) if not resolved_placeholder: logger.warning('No placeholder could be found for task name %s \ stage name %s and pipeline name %s. Please be sure to \ use object names and not uids in your references,i.e, \ $Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name)') raise ValueError( obj='placeholder', attribute='task', expected_value='$Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name) or $SHARED', actual_value=broken_placeholder) return resolved_placeholder except Exception, ex: logger.exception('Failed to resolve placeholder %s, error: %s' %(path, ex)) raise
python
def resolve_placeholders(path, placeholder_dict): """ **Purpose**: Substitute placeholders in staging attributes of a Task with actual paths to the corresponding tasks. :arguments: :path: string describing the staging paths, possibly containing a placeholder :placeholder_dict: dictionary holding the values for placeholders """ try: if isinstance(path, unicode): path = str(path) if not isinstance(path, str): raise TypeError(expected_type=str, actual_type=type(path)) if '$' not in path: return path # Extract placeholder from path if len(path.split('>')) == 1: placeholder = path.split('/')[0] else: if path.split('>')[0].strip().startswith('$'): placeholder = path.split('>')[0].strip().split('/')[0] else: placeholder = path.split('>')[1].strip().split('/')[0] # SHARED if placeholder == "$SHARED": return path.replace(placeholder, 'pilot://') # Expected placeholder format: # $Pipeline_{pipeline.uid}_Stage_{stage.uid}_Task_{task.uid} broken_placeholder = placeholder.split('/')[0].split('_') if not len(broken_placeholder) == 6: raise ValueError( obj='placeholder', attribute='task', expected_value='$Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name) or $SHARED', actual_value=broken_placeholder) pipeline_name = broken_placeholder[1] stage_name = broken_placeholder[3] task_name = broken_placeholder[5] resolved_placeholder = None if pipeline_name in placeholder_dict.keys(): if stage_name in placeholder_dict[pipeline_name].keys(): if task_name in placeholder_dict[pipeline_name][stage_name].keys(): resolved_placeholder = path.replace(placeholder, placeholder_dict[ pipeline_name][stage_name][task_name]['path']) else: logger.warning('%s not assigned to any task in Stage %s Pipeline %s' % (task_name, stage_name, pipeline_name)) else: logger.warning('%s not assigned to any Stage in Pipeline %s' % ( stage_name, pipeline_name)) else: logger.warning('%s not assigned to any Pipeline' % (pipeline_name)) if not resolved_placeholder: logger.warning('No placeholder could be found for task name %s \ stage name %s and pipeline name %s. Please be sure to \ use object names and not uids in your references,i.e, \ $Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name)') raise ValueError( obj='placeholder', attribute='task', expected_value='$Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name) or $SHARED', actual_value=broken_placeholder) return resolved_placeholder except Exception, ex: logger.exception('Failed to resolve placeholder %s, error: %s' %(path, ex)) raise
[ "def", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", ":", "try", ":", "if", "isinstance", "(", "path", ",", "unicode", ")", ":", "path", "=", "str", "(", "path", ")", "if", "not", "isinstance", "(", "path", ",", "str", ")", ":", "raise", "TypeError", "(", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "path", ")", ")", "if", "'$'", "not", "in", "path", ":", "return", "path", "# Extract placeholder from path", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", "==", "1", ":", "placeholder", "=", "path", ".", "split", "(", "'/'", ")", "[", "0", "]", "else", ":", "if", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ".", "startswith", "(", "'$'", ")", ":", "placeholder", "=", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ".", "split", "(", "'/'", ")", "[", "0", "]", "else", ":", "placeholder", "=", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", ".", "split", "(", "'/'", ")", "[", "0", "]", "# SHARED", "if", "placeholder", "==", "\"$SHARED\"", ":", "return", "path", ".", "replace", "(", "placeholder", ",", "'pilot://'", ")", "# Expected placeholder format:", "# $Pipeline_{pipeline.uid}_Stage_{stage.uid}_Task_{task.uid}", "broken_placeholder", "=", "placeholder", ".", "split", "(", "'/'", ")", "[", "0", "]", ".", "split", "(", "'_'", ")", "if", "not", "len", "(", "broken_placeholder", ")", "==", "6", ":", "raise", "ValueError", "(", "obj", "=", "'placeholder'", ",", "attribute", "=", "'task'", ",", "expected_value", "=", "'$Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name) or $SHARED'", ",", "actual_value", "=", "broken_placeholder", ")", "pipeline_name", "=", "broken_placeholder", "[", "1", "]", "stage_name", "=", "broken_placeholder", "[", "3", "]", "task_name", "=", "broken_placeholder", "[", "5", "]", "resolved_placeholder", "=", "None", "if", "pipeline_name", "in", "placeholder_dict", ".", "keys", "(", ")", ":", "if", "stage_name", "in", "placeholder_dict", "[", "pipeline_name", "]", ".", "keys", "(", ")", ":", "if", "task_name", "in", "placeholder_dict", "[", "pipeline_name", "]", "[", "stage_name", "]", ".", "keys", "(", ")", ":", "resolved_placeholder", "=", "path", ".", "replace", "(", "placeholder", ",", "placeholder_dict", "[", "pipeline_name", "]", "[", "stage_name", "]", "[", "task_name", "]", "[", "'path'", "]", ")", "else", ":", "logger", ".", "warning", "(", "'%s not assigned to any task in Stage %s Pipeline %s'", "%", "(", "task_name", ",", "stage_name", ",", "pipeline_name", ")", ")", "else", ":", "logger", ".", "warning", "(", "'%s not assigned to any Stage in Pipeline %s'", "%", "(", "stage_name", ",", "pipeline_name", ")", ")", "else", ":", "logger", ".", "warning", "(", "'%s not assigned to any Pipeline'", "%", "(", "pipeline_name", ")", ")", "if", "not", "resolved_placeholder", ":", "logger", ".", "warning", "(", "'No placeholder could be found for task name %s \\\n stage name %s and pipeline name %s. Please be sure to \\\n use object names and not uids in your references,i.e, \\\n $Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name)'", ")", "raise", "ValueError", "(", "obj", "=", "'placeholder'", ",", "attribute", "=", "'task'", ",", "expected_value", "=", "'$Pipeline_(pipeline_name)_Stage_(stage_name)_Task_(task_name) or $SHARED'", ",", "actual_value", "=", "broken_placeholder", ")", "return", "resolved_placeholder", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Failed to resolve placeholder %s, error: %s'", "%", "(", "path", ",", "ex", ")", ")", "raise" ]
**Purpose**: Substitute placeholders in staging attributes of a Task with actual paths to the corresponding tasks. :arguments: :path: string describing the staging paths, possibly containing a placeholder :placeholder_dict: dictionary holding the values for placeholders
[ "**", "Purpose", "**", ":", "Substitute", "placeholders", "in", "staging", "attributes", "of", "a", "Task", "with", "actual", "paths", "to", "the", "corresponding", "tasks", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/task_processor.py#L10-L91
radical-cybertools/radical.entk
src/radical/entk/execman/rp/task_processor.py
get_input_list_from_task
def get_input_list_from_task(task, placeholder_dict): """ Purpose: Parse a Task object to extract the files to be staged as the output. Details: The extracted data is then converted into the appropriate RP directive depending on whether the data is to be copied/downloaded. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: list of RP directives for the files that need to be staged out """ try: if not isinstance(task, Task): raise TypeError(expected_type=Task, actual_type=type(task)) input_data = [] if task.link_input_data: for path in task.link_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.LINK } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.LINK } input_data.append(temp) if task.upload_input_data: for path in task.upload_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip() } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()) } input_data.append(temp) if task.copy_input_data: for path in task.copy_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.COPY } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.COPY } input_data.append(temp) if task.move_input_data: for path in task.move_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.MOVE } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.MOVE } input_data.append(temp) return input_data except Exception, ex: logger.exception('Failed to get input list of files from task, error: %s' % ex) raise
python
def get_input_list_from_task(task, placeholder_dict): """ Purpose: Parse a Task object to extract the files to be staged as the output. Details: The extracted data is then converted into the appropriate RP directive depending on whether the data is to be copied/downloaded. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: list of RP directives for the files that need to be staged out """ try: if not isinstance(task, Task): raise TypeError(expected_type=Task, actual_type=type(task)) input_data = [] if task.link_input_data: for path in task.link_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.LINK } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.LINK } input_data.append(temp) if task.upload_input_data: for path in task.upload_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip() } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()) } input_data.append(temp) if task.copy_input_data: for path in task.copy_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.COPY } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.COPY } input_data.append(temp) if task.move_input_data: for path in task.move_input_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.MOVE } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.MOVE } input_data.append(temp) return input_data except Exception, ex: logger.exception('Failed to get input list of files from task, error: %s' % ex) raise
[ "def", "get_input_list_from_task", "(", "task", ",", "placeholder_dict", ")", ":", "try", ":", "if", "not", "isinstance", "(", "task", ",", "Task", ")", ":", "raise", "TypeError", "(", "expected_type", "=", "Task", ",", "actual_type", "=", "type", "(", "task", ")", ")", "input_data", "=", "[", "]", "if", "task", ".", "link_input_data", ":", "for", "path", "in", "task", ".", "link_input_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", ",", "'action'", ":", "rp", ".", "LINK", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", ",", "'action'", ":", "rp", ".", "LINK", "}", "input_data", ".", "append", "(", "temp", ")", "if", "task", ".", "upload_input_data", ":", "for", "path", "in", "task", ".", "upload_input_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "}", "input_data", ".", "append", "(", "temp", ")", "if", "task", ".", "copy_input_data", ":", "for", "path", "in", "task", ".", "copy_input_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", ",", "'action'", ":", "rp", ".", "COPY", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", ",", "'action'", ":", "rp", ".", "COPY", "}", "input_data", ".", "append", "(", "temp", ")", "if", "task", ".", "move_input_data", ":", "for", "path", "in", "task", ".", "move_input_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", ",", "'action'", ":", "rp", ".", "MOVE", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", ",", "'action'", ":", "rp", ".", "MOVE", "}", "input_data", ".", "append", "(", "temp", ")", "return", "input_data", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Failed to get input list of files from task, error: %s'", "%", "ex", ")", "raise" ]
Purpose: Parse a Task object to extract the files to be staged as the output. Details: The extracted data is then converted into the appropriate RP directive depending on whether the data is to be copied/downloaded. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: list of RP directives for the files that need to be staged out
[ "Purpose", ":", "Parse", "a", "Task", "object", "to", "extract", "the", "files", "to", "be", "staged", "as", "the", "output", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/task_processor.py#L156-L265
radical-cybertools/radical.entk
src/radical/entk/execman/rp/task_processor.py
get_output_list_from_task
def get_output_list_from_task(task, placeholder_dict): """ Purpose: Parse a Task object to extract the files to be staged as the output. Details: The extracted data is then converted into the appropriate RP directive depending on whether the data is to be copied/downloaded. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: list of RP directives for the files that need to be staged out """ try: if not isinstance(task, Task): raise TypeError(expected_type=Task, actual_type=type(task)) output_data = [] if task.copy_output_data: for path in task.copy_output_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.COPY } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.COPY } output_data.append(temp) if task.download_output_data: for path in task.download_output_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip() } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()) } output_data.append(temp) if task.move_output_data: for path in task.move_output_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.MOVE } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.MOVE } output_data.append(temp) return output_data except Exception, ex: logger.exception('Failed to get output list of files from task, error: %s' % ex) raise
python
def get_output_list_from_task(task, placeholder_dict): """ Purpose: Parse a Task object to extract the files to be staged as the output. Details: The extracted data is then converted into the appropriate RP directive depending on whether the data is to be copied/downloaded. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: list of RP directives for the files that need to be staged out """ try: if not isinstance(task, Task): raise TypeError(expected_type=Task, actual_type=type(task)) output_data = [] if task.copy_output_data: for path in task.copy_output_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.COPY } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.COPY } output_data.append(temp) if task.download_output_data: for path in task.download_output_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip() } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()) } output_data.append(temp) if task.move_output_data: for path in task.move_output_data: path = resolve_placeholders(path, placeholder_dict) if len(path.split('>')) > 1: temp = { 'source': path.split('>')[0].strip(), 'target': path.split('>')[1].strip(), 'action': rp.MOVE } else: temp = { 'source': path.split('>')[0].strip(), 'target': os.path.basename(path.split('>')[0].strip()), 'action': rp.MOVE } output_data.append(temp) return output_data except Exception, ex: logger.exception('Failed to get output list of files from task, error: %s' % ex) raise
[ "def", "get_output_list_from_task", "(", "task", ",", "placeholder_dict", ")", ":", "try", ":", "if", "not", "isinstance", "(", "task", ",", "Task", ")", ":", "raise", "TypeError", "(", "expected_type", "=", "Task", ",", "actual_type", "=", "type", "(", "task", ")", ")", "output_data", "=", "[", "]", "if", "task", ".", "copy_output_data", ":", "for", "path", "in", "task", ".", "copy_output_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", ",", "'action'", ":", "rp", ".", "COPY", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", ",", "'action'", ":", "rp", ".", "COPY", "}", "output_data", ".", "append", "(", "temp", ")", "if", "task", ".", "download_output_data", ":", "for", "path", "in", "task", ".", "download_output_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "}", "output_data", ".", "append", "(", "temp", ")", "if", "task", ".", "move_output_data", ":", "for", "path", "in", "task", ".", "move_output_data", ":", "path", "=", "resolve_placeholders", "(", "path", ",", "placeholder_dict", ")", "if", "len", "(", "path", ".", "split", "(", "'>'", ")", ")", ">", "1", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "path", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "strip", "(", ")", ",", "'action'", ":", "rp", ".", "MOVE", "}", "else", ":", "temp", "=", "{", "'source'", ":", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ",", "'target'", ":", "os", ".", "path", ".", "basename", "(", "path", ".", "split", "(", "'>'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", ",", "'action'", ":", "rp", ".", "MOVE", "}", "output_data", ".", "append", "(", "temp", ")", "return", "output_data", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Failed to get output list of files from task, error: %s'", "%", "ex", ")", "raise" ]
Purpose: Parse a Task object to extract the files to be staged as the output. Details: The extracted data is then converted into the appropriate RP directive depending on whether the data is to be copied/downloaded. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: list of RP directives for the files that need to be staged out
[ "Purpose", ":", "Parse", "a", "Task", "object", "to", "extract", "the", "files", "to", "be", "staged", "as", "the", "output", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/task_processor.py#L268-L357
radical-cybertools/radical.entk
src/radical/entk/execman/rp/task_processor.py
create_cud_from_task
def create_cud_from_task(task, placeholder_dict, prof=None): """ Purpose: Create a Compute Unit description based on the defined Task. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: ComputeUnitDescription """ try: logger.debug('Creating CU from Task %s' % (task.uid)) if prof: prof.prof('cud from task - create', uid=task.uid) cud = rp.ComputeUnitDescription() cud.name = '%s,%s,%s,%s,%s,%s' % (task.uid, task.name, task.parent_stage['uid'], task.parent_stage['name'], task.parent_pipeline['uid'], task.parent_pipeline['name']) cud.pre_exec = task.pre_exec cud.executable = task.executable cud.arguments = resolve_arguments(task.arguments, placeholder_dict) cud.post_exec = task.post_exec if task.tag: if task.parent_pipeline['name']: cud.tag = resolve_tags( tag=task.tag, parent_pipeline_name=task.parent_pipeline['name'], placeholder_dict=placeholder_dict) cud.cpu_processes = task.cpu_reqs['processes'] cud.cpu_threads = task.cpu_reqs['threads_per_process'] cud.cpu_process_type = task.cpu_reqs['process_type'] cud.cpu_thread_type = task.cpu_reqs['thread_type'] cud.gpu_processes = task.gpu_reqs['processes'] cud.gpu_threads = task.gpu_reqs['threads_per_process'] cud.gpu_process_type = task.gpu_reqs['process_type'] cud.gpu_thread_type = task.gpu_reqs['thread_type'] if task.lfs_per_process: cud.lfs_per_process = task.lfs_per_process if task.stdout: cud.stdout = task.stdout if task.stderr: cud.stderr = task.stderr cud.input_staging = get_input_list_from_task(task, placeholder_dict) cud.output_staging = get_output_list_from_task(task, placeholder_dict) if prof: prof.prof('cud from task - done', uid=task.uid) logger.debug('CU %s created from Task %s' % (cud.name, task.uid)) return cud except Exception, ex: logger.exception('CU creation failed, error: %s' % ex) raise
python
def create_cud_from_task(task, placeholder_dict, prof=None): """ Purpose: Create a Compute Unit description based on the defined Task. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: ComputeUnitDescription """ try: logger.debug('Creating CU from Task %s' % (task.uid)) if prof: prof.prof('cud from task - create', uid=task.uid) cud = rp.ComputeUnitDescription() cud.name = '%s,%s,%s,%s,%s,%s' % (task.uid, task.name, task.parent_stage['uid'], task.parent_stage['name'], task.parent_pipeline['uid'], task.parent_pipeline['name']) cud.pre_exec = task.pre_exec cud.executable = task.executable cud.arguments = resolve_arguments(task.arguments, placeholder_dict) cud.post_exec = task.post_exec if task.tag: if task.parent_pipeline['name']: cud.tag = resolve_tags( tag=task.tag, parent_pipeline_name=task.parent_pipeline['name'], placeholder_dict=placeholder_dict) cud.cpu_processes = task.cpu_reqs['processes'] cud.cpu_threads = task.cpu_reqs['threads_per_process'] cud.cpu_process_type = task.cpu_reqs['process_type'] cud.cpu_thread_type = task.cpu_reqs['thread_type'] cud.gpu_processes = task.gpu_reqs['processes'] cud.gpu_threads = task.gpu_reqs['threads_per_process'] cud.gpu_process_type = task.gpu_reqs['process_type'] cud.gpu_thread_type = task.gpu_reqs['thread_type'] if task.lfs_per_process: cud.lfs_per_process = task.lfs_per_process if task.stdout: cud.stdout = task.stdout if task.stderr: cud.stderr = task.stderr cud.input_staging = get_input_list_from_task(task, placeholder_dict) cud.output_staging = get_output_list_from_task(task, placeholder_dict) if prof: prof.prof('cud from task - done', uid=task.uid) logger.debug('CU %s created from Task %s' % (cud.name, task.uid)) return cud except Exception, ex: logger.exception('CU creation failed, error: %s' % ex) raise
[ "def", "create_cud_from_task", "(", "task", ",", "placeholder_dict", ",", "prof", "=", "None", ")", ":", "try", ":", "logger", ".", "debug", "(", "'Creating CU from Task %s'", "%", "(", "task", ".", "uid", ")", ")", "if", "prof", ":", "prof", ".", "prof", "(", "'cud from task - create'", ",", "uid", "=", "task", ".", "uid", ")", "cud", "=", "rp", ".", "ComputeUnitDescription", "(", ")", "cud", ".", "name", "=", "'%s,%s,%s,%s,%s,%s'", "%", "(", "task", ".", "uid", ",", "task", ".", "name", ",", "task", ".", "parent_stage", "[", "'uid'", "]", ",", "task", ".", "parent_stage", "[", "'name'", "]", ",", "task", ".", "parent_pipeline", "[", "'uid'", "]", ",", "task", ".", "parent_pipeline", "[", "'name'", "]", ")", "cud", ".", "pre_exec", "=", "task", ".", "pre_exec", "cud", ".", "executable", "=", "task", ".", "executable", "cud", ".", "arguments", "=", "resolve_arguments", "(", "task", ".", "arguments", ",", "placeholder_dict", ")", "cud", ".", "post_exec", "=", "task", ".", "post_exec", "if", "task", ".", "tag", ":", "if", "task", ".", "parent_pipeline", "[", "'name'", "]", ":", "cud", ".", "tag", "=", "resolve_tags", "(", "tag", "=", "task", ".", "tag", ",", "parent_pipeline_name", "=", "task", ".", "parent_pipeline", "[", "'name'", "]", ",", "placeholder_dict", "=", "placeholder_dict", ")", "cud", ".", "cpu_processes", "=", "task", ".", "cpu_reqs", "[", "'processes'", "]", "cud", ".", "cpu_threads", "=", "task", ".", "cpu_reqs", "[", "'threads_per_process'", "]", "cud", ".", "cpu_process_type", "=", "task", ".", "cpu_reqs", "[", "'process_type'", "]", "cud", ".", "cpu_thread_type", "=", "task", ".", "cpu_reqs", "[", "'thread_type'", "]", "cud", ".", "gpu_processes", "=", "task", ".", "gpu_reqs", "[", "'processes'", "]", "cud", ".", "gpu_threads", "=", "task", ".", "gpu_reqs", "[", "'threads_per_process'", "]", "cud", ".", "gpu_process_type", "=", "task", ".", "gpu_reqs", "[", "'process_type'", "]", "cud", ".", "gpu_thread_type", "=", "task", ".", "gpu_reqs", "[", "'thread_type'", "]", "if", "task", ".", "lfs_per_process", ":", "cud", ".", "lfs_per_process", "=", "task", ".", "lfs_per_process", "if", "task", ".", "stdout", ":", "cud", ".", "stdout", "=", "task", ".", "stdout", "if", "task", ".", "stderr", ":", "cud", ".", "stderr", "=", "task", ".", "stderr", "cud", ".", "input_staging", "=", "get_input_list_from_task", "(", "task", ",", "placeholder_dict", ")", "cud", ".", "output_staging", "=", "get_output_list_from_task", "(", "task", ",", "placeholder_dict", ")", "if", "prof", ":", "prof", ".", "prof", "(", "'cud from task - done'", ",", "uid", "=", "task", ".", "uid", ")", "logger", ".", "debug", "(", "'CU %s created from Task %s'", "%", "(", "cud", ".", "name", ",", "task", ".", "uid", ")", ")", "return", "cud", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'CU creation failed, error: %s'", "%", "ex", ")", "raise" ]
Purpose: Create a Compute Unit description based on the defined Task. :arguments: :task: EnTK Task object :placeholder_dict: dictionary holding the values for placeholders :return: ComputeUnitDescription
[ "Purpose", ":", "Create", "a", "Compute", "Unit", "description", "based", "on", "the", "defined", "Task", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/task_processor.py#L360-L420
radical-cybertools/radical.entk
src/radical/entk/execman/rp/task_processor.py
create_task_from_cu
def create_task_from_cu(cu, prof=None): """ Purpose: Create a Task based on the Compute Unit. Details: Currently, only the uid, parent_stage and parent_pipeline are retrieved. The exact initial Task (that was converted to a CUD) cannot be recovered as the RP API does not provide the same attributes for a CU as for a CUD. Also, this is not required for the most part. TODO: Add exit code, stdout, stderr and path attributes to a Task. These can be extracted from a CU :arguments: :cu: RP Compute Unit :return: Task """ try: logger.debug('Create Task from CU %s' % cu.name) if prof: prof.prof('task from cu - create', uid=cu.name.split(',')[0].strip()) task = Task() task.uid = cu.name.split(',')[0].strip() task.name = cu.name.split(',')[1].strip() task.parent_stage['uid'] = cu.name.split(',')[2].strip() task.parent_stage['name'] = cu.name.split(',')[3].strip() task.parent_pipeline['uid'] = cu.name.split(',')[4].strip() task.parent_pipeline['name'] = cu.name.split(',')[5].strip() task.rts_uid = cu.uid if cu.state == rp.DONE: task.exit_code = 0 else: task.exit_code = 1 task.path = ru.Url(cu.sandbox).path if prof: prof.prof('task from cu - done', uid=cu.name.split(',')[0].strip()) logger.debug('Task %s created from CU %s' % (task.uid, cu.name)) return task except Exception, ex: logger.exception('Task creation from CU failed, error: %s' % ex) raise
python
def create_task_from_cu(cu, prof=None): """ Purpose: Create a Task based on the Compute Unit. Details: Currently, only the uid, parent_stage and parent_pipeline are retrieved. The exact initial Task (that was converted to a CUD) cannot be recovered as the RP API does not provide the same attributes for a CU as for a CUD. Also, this is not required for the most part. TODO: Add exit code, stdout, stderr and path attributes to a Task. These can be extracted from a CU :arguments: :cu: RP Compute Unit :return: Task """ try: logger.debug('Create Task from CU %s' % cu.name) if prof: prof.prof('task from cu - create', uid=cu.name.split(',')[0].strip()) task = Task() task.uid = cu.name.split(',')[0].strip() task.name = cu.name.split(',')[1].strip() task.parent_stage['uid'] = cu.name.split(',')[2].strip() task.parent_stage['name'] = cu.name.split(',')[3].strip() task.parent_pipeline['uid'] = cu.name.split(',')[4].strip() task.parent_pipeline['name'] = cu.name.split(',')[5].strip() task.rts_uid = cu.uid if cu.state == rp.DONE: task.exit_code = 0 else: task.exit_code = 1 task.path = ru.Url(cu.sandbox).path if prof: prof.prof('task from cu - done', uid=cu.name.split(',')[0].strip()) logger.debug('Task %s created from CU %s' % (task.uid, cu.name)) return task except Exception, ex: logger.exception('Task creation from CU failed, error: %s' % ex) raise
[ "def", "create_task_from_cu", "(", "cu", ",", "prof", "=", "None", ")", ":", "try", ":", "logger", ".", "debug", "(", "'Create Task from CU %s'", "%", "cu", ".", "name", ")", "if", "prof", ":", "prof", ".", "prof", "(", "'task from cu - create'", ",", "uid", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "task", "=", "Task", "(", ")", "task", ".", "uid", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "0", "]", ".", "strip", "(", ")", "task", ".", "name", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "1", "]", ".", "strip", "(", ")", "task", ".", "parent_stage", "[", "'uid'", "]", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "2", "]", ".", "strip", "(", ")", "task", ".", "parent_stage", "[", "'name'", "]", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "3", "]", ".", "strip", "(", ")", "task", ".", "parent_pipeline", "[", "'uid'", "]", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "4", "]", ".", "strip", "(", ")", "task", ".", "parent_pipeline", "[", "'name'", "]", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "5", "]", ".", "strip", "(", ")", "task", ".", "rts_uid", "=", "cu", ".", "uid", "if", "cu", ".", "state", "==", "rp", ".", "DONE", ":", "task", ".", "exit_code", "=", "0", "else", ":", "task", ".", "exit_code", "=", "1", "task", ".", "path", "=", "ru", ".", "Url", "(", "cu", ".", "sandbox", ")", ".", "path", "if", "prof", ":", "prof", ".", "prof", "(", "'task from cu - done'", ",", "uid", "=", "cu", ".", "name", ".", "split", "(", "','", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "logger", ".", "debug", "(", "'Task %s created from CU %s'", "%", "(", "task", ".", "uid", ",", "cu", ".", "name", ")", ")", "return", "task", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Task creation from CU failed, error: %s'", "%", "ex", ")", "raise" ]
Purpose: Create a Task based on the Compute Unit. Details: Currently, only the uid, parent_stage and parent_pipeline are retrieved. The exact initial Task (that was converted to a CUD) cannot be recovered as the RP API does not provide the same attributes for a CU as for a CUD. Also, this is not required for the most part. TODO: Add exit code, stdout, stderr and path attributes to a Task. These can be extracted from a CU :arguments: :cu: RP Compute Unit :return: Task
[ "Purpose", ":", "Create", "a", "Task", "based", "on", "the", "Compute", "Unit", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/task_processor.py#L423-L472
bradmontgomery/django-redis-metrics
redis_metrics/management/commands/redis_metrics_send_mail.py
Command.handle_noargs
def handle_noargs(self, **options): """Send Report E-mails.""" r = get_r() since = datetime.utcnow() - timedelta(days=1) metrics = {} categories = r.metric_slugs_by_category() for category_name, slug_list in categories.items(): metrics[category_name] = [] for slug in slug_list: metric_values = r.get_metric_history(slug, since=since) metrics[category_name].append( (slug, metric_values) ) # metrics is now: # -------------- # { Category : [ # ('foo', [('m:foo:2012-07-18', 1), ('m:foo:2012-07-19, 2), ...]) # ], # ... # } template = "redis_metrics/email/report.{fmt}" data = { 'today': since, 'metrics': metrics, } message = render_to_string(template.format(fmt='txt'), data) message_html = render_to_string(template.format(fmt='html'), data) msg = EmailMultiAlternatives( subject="Redis Metrics Report", body=message, from_email=settings.DEFAULT_FROM_EMAIL, to=[email for name, email in settings.ADMINS] ) msg.attach_alternative(message_html, "text/html") msg.send()
python
def handle_noargs(self, **options): """Send Report E-mails.""" r = get_r() since = datetime.utcnow() - timedelta(days=1) metrics = {} categories = r.metric_slugs_by_category() for category_name, slug_list in categories.items(): metrics[category_name] = [] for slug in slug_list: metric_values = r.get_metric_history(slug, since=since) metrics[category_name].append( (slug, metric_values) ) # metrics is now: # -------------- # { Category : [ # ('foo', [('m:foo:2012-07-18', 1), ('m:foo:2012-07-19, 2), ...]) # ], # ... # } template = "redis_metrics/email/report.{fmt}" data = { 'today': since, 'metrics': metrics, } message = render_to_string(template.format(fmt='txt'), data) message_html = render_to_string(template.format(fmt='html'), data) msg = EmailMultiAlternatives( subject="Redis Metrics Report", body=message, from_email=settings.DEFAULT_FROM_EMAIL, to=[email for name, email in settings.ADMINS] ) msg.attach_alternative(message_html, "text/html") msg.send()
[ "def", "handle_noargs", "(", "self", ",", "*", "*", "options", ")", ":", "r", "=", "get_r", "(", ")", "since", "=", "datetime", ".", "utcnow", "(", ")", "-", "timedelta", "(", "days", "=", "1", ")", "metrics", "=", "{", "}", "categories", "=", "r", ".", "metric_slugs_by_category", "(", ")", "for", "category_name", ",", "slug_list", "in", "categories", ".", "items", "(", ")", ":", "metrics", "[", "category_name", "]", "=", "[", "]", "for", "slug", "in", "slug_list", ":", "metric_values", "=", "r", ".", "get_metric_history", "(", "slug", ",", "since", "=", "since", ")", "metrics", "[", "category_name", "]", ".", "append", "(", "(", "slug", ",", "metric_values", ")", ")", "# metrics is now:", "# --------------", "# { Category : [", "# ('foo', [('m:foo:2012-07-18', 1), ('m:foo:2012-07-19, 2), ...])", "# ],", "# ...", "# }", "template", "=", "\"redis_metrics/email/report.{fmt}\"", "data", "=", "{", "'today'", ":", "since", ",", "'metrics'", ":", "metrics", ",", "}", "message", "=", "render_to_string", "(", "template", ".", "format", "(", "fmt", "=", "'txt'", ")", ",", "data", ")", "message_html", "=", "render_to_string", "(", "template", ".", "format", "(", "fmt", "=", "'html'", ")", ",", "data", ")", "msg", "=", "EmailMultiAlternatives", "(", "subject", "=", "\"Redis Metrics Report\"", ",", "body", "=", "message", ",", "from_email", "=", "settings", ".", "DEFAULT_FROM_EMAIL", ",", "to", "=", "[", "email", "for", "name", ",", "email", "in", "settings", ".", "ADMINS", "]", ")", "msg", ".", "attach_alternative", "(", "message_html", ",", "\"text/html\"", ")", "msg", ".", "send", "(", ")" ]
Send Report E-mails.
[ "Send", "Report", "E", "-", "mails", "." ]
train
https://github.com/bradmontgomery/django-redis-metrics/blob/2c92332920113d28c39234b949aa496b39a091d1/redis_metrics/management/commands/redis_metrics_send_mail.py#L16-L54
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage.luid
def luid(self): """ Unique ID of the current stage (fully qualified). example: >>> stage.luid pipe.0001.stage.0004 :getter: Returns the fully qualified uid of the current stage :type: String """ p_elem = self.parent_pipeline.get('name') if not p_elem: p_elem = self.parent_pipeline['uid'] s_elem = self.name if not s_elem: s_elem = self.uid return '%s.%s' % (p_elem, s_elem)
python
def luid(self): """ Unique ID of the current stage (fully qualified). example: >>> stage.luid pipe.0001.stage.0004 :getter: Returns the fully qualified uid of the current stage :type: String """ p_elem = self.parent_pipeline.get('name') if not p_elem: p_elem = self.parent_pipeline['uid'] s_elem = self.name if not s_elem: s_elem = self.uid return '%s.%s' % (p_elem, s_elem)
[ "def", "luid", "(", "self", ")", ":", "p_elem", "=", "self", ".", "parent_pipeline", ".", "get", "(", "'name'", ")", "if", "not", "p_elem", ":", "p_elem", "=", "self", ".", "parent_pipeline", "[", "'uid'", "]", "s_elem", "=", "self", ".", "name", "if", "not", "s_elem", ":", "s_elem", "=", "self", ".", "uid", "return", "'%s.%s'", "%", "(", "p_elem", ",", "s_elem", ")" ]
Unique ID of the current stage (fully qualified). example: >>> stage.luid pipe.0001.stage.0004 :getter: Returns the fully qualified uid of the current stage :type: String
[ "Unique", "ID", "of", "the", "current", "stage", "(", "fully", "qualified", ")", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L92-L111
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage.add_tasks
def add_tasks(self, value): """ Adds tasks to the existing set of tasks of the Stage :argument: set of tasks """ tasks = self._validate_entities(value) self._tasks.update(tasks) self._task_count = len(self._tasks)
python
def add_tasks(self, value): """ Adds tasks to the existing set of tasks of the Stage :argument: set of tasks """ tasks = self._validate_entities(value) self._tasks.update(tasks) self._task_count = len(self._tasks)
[ "def", "add_tasks", "(", "self", ",", "value", ")", ":", "tasks", "=", "self", ".", "_validate_entities", "(", "value", ")", "self", ".", "_tasks", ".", "update", "(", "tasks", ")", "self", ".", "_task_count", "=", "len", "(", "self", ".", "_tasks", ")" ]
Adds tasks to the existing set of tasks of the Stage :argument: set of tasks
[ "Adds", "tasks", "to", "the", "existing", "set", "of", "tasks", "of", "the", "Stage" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L201-L209
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage.to_dict
def to_dict(self): """ Convert current Stage into a dictionary :return: python dictionary """ stage_desc_as_dict = { 'uid': self._uid, 'name': self._name, 'state': self._state, 'state_history': self._state_history, 'parent_pipeline': self._p_pipeline } return stage_desc_as_dict
python
def to_dict(self): """ Convert current Stage into a dictionary :return: python dictionary """ stage_desc_as_dict = { 'uid': self._uid, 'name': self._name, 'state': self._state, 'state_history': self._state_history, 'parent_pipeline': self._p_pipeline } return stage_desc_as_dict
[ "def", "to_dict", "(", "self", ")", ":", "stage_desc_as_dict", "=", "{", "'uid'", ":", "self", ".", "_uid", ",", "'name'", ":", "self", ".", "_name", ",", "'state'", ":", "self", ".", "_state", ",", "'state_history'", ":", "self", ".", "_state_history", ",", "'parent_pipeline'", ":", "self", ".", "_p_pipeline", "}", "return", "stage_desc_as_dict" ]
Convert current Stage into a dictionary :return: python dictionary
[ "Convert", "current", "Stage", "into", "a", "dictionary" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L211-L227
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage.from_dict
def from_dict(self, d): """ Create a Stage from a dictionary. The change is in inplace. :argument: python dictionary :return: None """ if 'uid' in d: if d['uid']: self._uid = d['uid'] if 'name' in d: if d['name']: self._name = d['name'] if 'state' in d: if isinstance(d['state'], str) or isinstance(d['state'], unicode): if d['state'] in states._stage_state_values.keys(): self._state = d['state'] else: raise ValueError(obj=self._uid, attribute='state', expected_value=states._stage_state_values.keys(), actual_value=value) else: raise TypeError(entity='state', expected_type=str, actual_type=type(d['state'])) else: self._state = states.INITIAL if 'state_history' in d: if isinstance(d['state_history'], list): self._state_history = d['state_history'] else: raise TypeError(entity='state_history', expected_type=list, actual_type=type(d['state_history'])) if 'parent_pipeline' in d: if isinstance(d['parent_pipeline'], dict): self._p_pipeline = d['parent_pipeline'] else: raise TypeError(entity='parent_pipeline', expected_type=dict, actual_type=type(d['parent_pipeline']))
python
def from_dict(self, d): """ Create a Stage from a dictionary. The change is in inplace. :argument: python dictionary :return: None """ if 'uid' in d: if d['uid']: self._uid = d['uid'] if 'name' in d: if d['name']: self._name = d['name'] if 'state' in d: if isinstance(d['state'], str) or isinstance(d['state'], unicode): if d['state'] in states._stage_state_values.keys(): self._state = d['state'] else: raise ValueError(obj=self._uid, attribute='state', expected_value=states._stage_state_values.keys(), actual_value=value) else: raise TypeError(entity='state', expected_type=str, actual_type=type(d['state'])) else: self._state = states.INITIAL if 'state_history' in d: if isinstance(d['state_history'], list): self._state_history = d['state_history'] else: raise TypeError(entity='state_history', expected_type=list, actual_type=type(d['state_history'])) if 'parent_pipeline' in d: if isinstance(d['parent_pipeline'], dict): self._p_pipeline = d['parent_pipeline'] else: raise TypeError(entity='parent_pipeline', expected_type=dict, actual_type=type(d['parent_pipeline']))
[ "def", "from_dict", "(", "self", ",", "d", ")", ":", "if", "'uid'", "in", "d", ":", "if", "d", "[", "'uid'", "]", ":", "self", ".", "_uid", "=", "d", "[", "'uid'", "]", "if", "'name'", "in", "d", ":", "if", "d", "[", "'name'", "]", ":", "self", ".", "_name", "=", "d", "[", "'name'", "]", "if", "'state'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'state'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'state'", "]", ",", "unicode", ")", ":", "if", "d", "[", "'state'", "]", "in", "states", ".", "_stage_state_values", ".", "keys", "(", ")", ":", "self", ".", "_state", "=", "d", "[", "'state'", "]", "else", ":", "raise", "ValueError", "(", "obj", "=", "self", ".", "_uid", ",", "attribute", "=", "'state'", ",", "expected_value", "=", "states", ".", "_stage_state_values", ".", "keys", "(", ")", ",", "actual_value", "=", "value", ")", "else", ":", "raise", "TypeError", "(", "entity", "=", "'state'", ",", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'state'", "]", ")", ")", "else", ":", "self", ".", "_state", "=", "states", ".", "INITIAL", "if", "'state_history'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'state_history'", "]", ",", "list", ")", ":", "self", ".", "_state_history", "=", "d", "[", "'state_history'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'state_history'", ",", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'state_history'", "]", ")", ")", "if", "'parent_pipeline'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'parent_pipeline'", "]", ",", "dict", ")", ":", "self", ".", "_p_pipeline", "=", "d", "[", "'parent_pipeline'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'parent_pipeline'", ",", "expected_type", "=", "dict", ",", "actual_type", "=", "type", "(", "d", "[", "'parent_pipeline'", "]", ")", ")" ]
Create a Stage from a dictionary. The change is in inplace. :argument: python dictionary :return: None
[ "Create", "a", "Stage", "from", "a", "dictionary", ".", "The", "change", "is", "in", "inplace", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L229-L270
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage._set_tasks_state
def _set_tasks_state(self, value): """ Purpose: Set state of all tasks of the current stage. :arguments: String """ if value not in states.state_numbers.keys(): raise ValueError(obj=self._uid, attribute='set_tasks_state', expected_value=states.state_numbers.keys(), actual_value=value) for task in self._tasks: task.state = value
python
def _set_tasks_state(self, value): """ Purpose: Set state of all tasks of the current stage. :arguments: String """ if value not in states.state_numbers.keys(): raise ValueError(obj=self._uid, attribute='set_tasks_state', expected_value=states.state_numbers.keys(), actual_value=value) for task in self._tasks: task.state = value
[ "def", "_set_tasks_state", "(", "self", ",", "value", ")", ":", "if", "value", "not", "in", "states", ".", "state_numbers", ".", "keys", "(", ")", ":", "raise", "ValueError", "(", "obj", "=", "self", ".", "_uid", ",", "attribute", "=", "'set_tasks_state'", ",", "expected_value", "=", "states", ".", "state_numbers", ".", "keys", "(", ")", ",", "actual_value", "=", "value", ")", "for", "task", "in", "self", ".", "_tasks", ":", "task", ".", "state", "=", "value" ]
Purpose: Set state of all tasks of the current stage. :arguments: String
[ "Purpose", ":", "Set", "state", "of", "all", "tasks", "of", "the", "current", "stage", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L276-L289
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage._check_stage_complete
def _check_stage_complete(self): """ Purpose: Check if all tasks of the current stage have completed, i.e., are in either DONE or FAILED state. """ try: for task in self._tasks: if task.state not in [states.DONE, states.FAILED]: return False return True except Exception, ex: raise EnTKError(ex)
python
def _check_stage_complete(self): """ Purpose: Check if all tasks of the current stage have completed, i.e., are in either DONE or FAILED state. """ try: for task in self._tasks: if task.state not in [states.DONE, states.FAILED]: return False return True except Exception, ex: raise EnTKError(ex)
[ "def", "_check_stage_complete", "(", "self", ")", ":", "try", ":", "for", "task", "in", "self", ".", "_tasks", ":", "if", "task", ".", "state", "not", "in", "[", "states", ".", "DONE", ",", "states", ".", "FAILED", "]", ":", "return", "False", "return", "True", "except", "Exception", ",", "ex", ":", "raise", "EnTKError", "(", "ex", ")" ]
Purpose: Check if all tasks of the current stage have completed, i.e., are in either DONE or FAILED state.
[ "Purpose", ":", "Check", "if", "all", "tasks", "of", "the", "current", "stage", "have", "completed", "i", ".", "e", ".", "are", "in", "either", "DONE", "or", "FAILED", "state", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L291-L305
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage._validate_entities
def _validate_entities(self, tasks): """ Purpose: Validate whether the 'tasks' is of type set. Validate the description of each Task. """ if not tasks: raise TypeError(expected_type=Task, actual_type=type(tasks)) if not isinstance(tasks, set): if not isinstance(tasks, list): tasks = set([tasks]) else: tasks = set(tasks) for t in tasks: if not isinstance(t, Task): raise TypeError(expected_type=Task, actual_type=type(t)) return tasks
python
def _validate_entities(self, tasks): """ Purpose: Validate whether the 'tasks' is of type set. Validate the description of each Task. """ if not tasks: raise TypeError(expected_type=Task, actual_type=type(tasks)) if not isinstance(tasks, set): if not isinstance(tasks, list): tasks = set([tasks]) else: tasks = set(tasks) for t in tasks: if not isinstance(t, Task): raise TypeError(expected_type=Task, actual_type=type(t)) return tasks
[ "def", "_validate_entities", "(", "self", ",", "tasks", ")", ":", "if", "not", "tasks", ":", "raise", "TypeError", "(", "expected_type", "=", "Task", ",", "actual_type", "=", "type", "(", "tasks", ")", ")", "if", "not", "isinstance", "(", "tasks", ",", "set", ")", ":", "if", "not", "isinstance", "(", "tasks", ",", "list", ")", ":", "tasks", "=", "set", "(", "[", "tasks", "]", ")", "else", ":", "tasks", "=", "set", "(", "tasks", ")", "for", "t", "in", "tasks", ":", "if", "not", "isinstance", "(", "t", ",", "Task", ")", ":", "raise", "TypeError", "(", "expected_type", "=", "Task", ",", "actual_type", "=", "type", "(", "t", ")", ")", "return", "tasks" ]
Purpose: Validate whether the 'tasks' is of type set. Validate the description of each Task.
[ "Purpose", ":", "Validate", "whether", "the", "tasks", "is", "of", "type", "set", ".", "Validate", "the", "description", "of", "each", "Task", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L308-L328
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage._assign_uid
def _assign_uid(self, sid): """ Purpose: Assign a uid to the current object based on the sid passed. Pass the current uid to children of current object """ self._uid = ru.generate_id('stage.%(item_counter)04d', ru.ID_CUSTOM, namespace=sid) for task in self._tasks: task._assign_uid(sid) self._pass_uid()
python
def _assign_uid(self, sid): """ Purpose: Assign a uid to the current object based on the sid passed. Pass the current uid to children of current object """ self._uid = ru.generate_id('stage.%(item_counter)04d', ru.ID_CUSTOM, namespace=sid) for task in self._tasks: task._assign_uid(sid) self._pass_uid()
[ "def", "_assign_uid", "(", "self", ",", "sid", ")", ":", "self", ".", "_uid", "=", "ru", ".", "generate_id", "(", "'stage.%(item_counter)04d'", ",", "ru", ".", "ID_CUSTOM", ",", "namespace", "=", "sid", ")", "for", "task", "in", "self", ".", "_tasks", ":", "task", ".", "_assign_uid", "(", "sid", ")", "self", ".", "_pass_uid", "(", ")" ]
Purpose: Assign a uid to the current object based on the sid passed. Pass the current uid to children of current object
[ "Purpose", ":", "Assign", "a", "uid", "to", "the", "current", "object", "based", "on", "the", "sid", "passed", ".", "Pass", "the", "current", "uid", "to", "children", "of", "current", "object" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L351-L360
radical-cybertools/radical.entk
src/radical/entk/stage/stage.py
Stage._pass_uid
def _pass_uid(self): """ Purpose: Assign the parent Stage and the parent Pipeline to all the tasks of the current stage. :arguments: set of Tasks (optional) :return: list of updated Tasks """ for task in self._tasks: task.parent_stage['uid'] = self._uid task.parent_stage['name'] = self._name task.parent_pipeline['uid'] = self._p_pipeline['uid'] task.parent_pipeline['name'] = self._p_pipeline['name']
python
def _pass_uid(self): """ Purpose: Assign the parent Stage and the parent Pipeline to all the tasks of the current stage. :arguments: set of Tasks (optional) :return: list of updated Tasks """ for task in self._tasks: task.parent_stage['uid'] = self._uid task.parent_stage['name'] = self._name task.parent_pipeline['uid'] = self._p_pipeline['uid'] task.parent_pipeline['name'] = self._p_pipeline['name']
[ "def", "_pass_uid", "(", "self", ")", ":", "for", "task", "in", "self", ".", "_tasks", ":", "task", ".", "parent_stage", "[", "'uid'", "]", "=", "self", ".", "_uid", "task", ".", "parent_stage", "[", "'name'", "]", "=", "self", ".", "_name", "task", ".", "parent_pipeline", "[", "'uid'", "]", "=", "self", ".", "_p_pipeline", "[", "'uid'", "]", "task", ".", "parent_pipeline", "[", "'name'", "]", "=", "self", ".", "_p_pipeline", "[", "'name'", "]" ]
Purpose: Assign the parent Stage and the parent Pipeline to all the tasks of the current stage. :arguments: set of Tasks (optional) :return: list of updated Tasks
[ "Purpose", ":", "Assign", "the", "parent", "Stage", "and", "the", "parent", "Pipeline", "to", "all", "the", "tasks", "of", "the", "current", "stage", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/stage/stage.py#L362-L374
google/dotty
efilter/parsers/dottysql/grammar.py
application
def application(tokens): """Matches function call (application).""" tokens = iter(tokens) func = next(tokens) paren = next(tokens) if func and func.name == "symbol" and paren.name == "lparen": # We would be able to unambiguously parse function application with # whitespace between the function name and the lparen, but let's not # do that because it's unexpected in most languages. if func.end != paren.start: raise errors.EfilterParseError( start=func.start, end=paren.end, message="No whitespace allowed between function and paren.") return common.TokenMatch(None, func.value, (func, paren))
python
def application(tokens): """Matches function call (application).""" tokens = iter(tokens) func = next(tokens) paren = next(tokens) if func and func.name == "symbol" and paren.name == "lparen": # We would be able to unambiguously parse function application with # whitespace between the function name and the lparen, but let's not # do that because it's unexpected in most languages. if func.end != paren.start: raise errors.EfilterParseError( start=func.start, end=paren.end, message="No whitespace allowed between function and paren.") return common.TokenMatch(None, func.value, (func, paren))
[ "def", "application", "(", "tokens", ")", ":", "tokens", "=", "iter", "(", "tokens", ")", "func", "=", "next", "(", "tokens", ")", "paren", "=", "next", "(", "tokens", ")", "if", "func", "and", "func", ".", "name", "==", "\"symbol\"", "and", "paren", ".", "name", "==", "\"lparen\"", ":", "# We would be able to unambiguously parse function application with", "# whitespace between the function name and the lparen, but let's not", "# do that because it's unexpected in most languages.", "if", "func", ".", "end", "!=", "paren", ".", "start", ":", "raise", "errors", ".", "EfilterParseError", "(", "start", "=", "func", ".", "start", ",", "end", "=", "paren", ".", "end", ",", "message", "=", "\"No whitespace allowed between function and paren.\"", ")", "return", "common", ".", "TokenMatch", "(", "None", ",", "func", ".", "value", ",", "(", "func", ",", "paren", ")", ")" ]
Matches function call (application).
[ "Matches", "function", "call", "(", "application", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/dottysql/grammar.py#L208-L223
google/dotty
setup.py
BdistRPMCommand._make_spec_file
def _make_spec_file(self): """Generates the text of an RPM spec file. Returns: A list of strings containing the lines of text. """ # Note that bdist_rpm can be an old style class. if issubclass(BdistRPMCommand, object): spec_file = super(BdistRPMCommand, self)._make_spec_file() else: spec_file = bdist_rpm._make_spec_file(self) if sys.version_info[0] < 3: python_package = "python" else: python_package = "python3" description = [] summary = "" in_description = False python_spec_file = [] for line in spec_file: if line.startswith("Summary: "): summary = line elif line.startswith("BuildRequires: "): line = "BuildRequires: {0:s}-setuptools".format(python_package) elif line.startswith("Requires: "): if python_package == "python3": line = line.replace("python", "python3") elif line.startswith("%description"): in_description = True elif line.startswith("%files"): line = "%files -f INSTALLED_FILES -n {0:s}-%{{name}}".format( python_package) elif line.startswith("%prep"): in_description = False python_spec_file.append( "%package -n {0:s}-%{{name}}".format(python_package)) python_spec_file.append("{0:s}".format(summary)) python_spec_file.append("") python_spec_file.append( "%description -n {0:s}-%{{name}}".format(python_package)) python_spec_file.extend(description) elif in_description: # Ignore leading white lines in the description. if not description and not line: continue description.append(line) python_spec_file.append(line) return python_spec_file
python
def _make_spec_file(self): """Generates the text of an RPM spec file. Returns: A list of strings containing the lines of text. """ # Note that bdist_rpm can be an old style class. if issubclass(BdistRPMCommand, object): spec_file = super(BdistRPMCommand, self)._make_spec_file() else: spec_file = bdist_rpm._make_spec_file(self) if sys.version_info[0] < 3: python_package = "python" else: python_package = "python3" description = [] summary = "" in_description = False python_spec_file = [] for line in spec_file: if line.startswith("Summary: "): summary = line elif line.startswith("BuildRequires: "): line = "BuildRequires: {0:s}-setuptools".format(python_package) elif line.startswith("Requires: "): if python_package == "python3": line = line.replace("python", "python3") elif line.startswith("%description"): in_description = True elif line.startswith("%files"): line = "%files -f INSTALLED_FILES -n {0:s}-%{{name}}".format( python_package) elif line.startswith("%prep"): in_description = False python_spec_file.append( "%package -n {0:s}-%{{name}}".format(python_package)) python_spec_file.append("{0:s}".format(summary)) python_spec_file.append("") python_spec_file.append( "%description -n {0:s}-%{{name}}".format(python_package)) python_spec_file.extend(description) elif in_description: # Ignore leading white lines in the description. if not description and not line: continue description.append(line) python_spec_file.append(line) return python_spec_file
[ "def", "_make_spec_file", "(", "self", ")", ":", "# Note that bdist_rpm can be an old style class.", "if", "issubclass", "(", "BdistRPMCommand", ",", "object", ")", ":", "spec_file", "=", "super", "(", "BdistRPMCommand", ",", "self", ")", ".", "_make_spec_file", "(", ")", "else", ":", "spec_file", "=", "bdist_rpm", ".", "_make_spec_file", "(", "self", ")", "if", "sys", ".", "version_info", "[", "0", "]", "<", "3", ":", "python_package", "=", "\"python\"", "else", ":", "python_package", "=", "\"python3\"", "description", "=", "[", "]", "summary", "=", "\"\"", "in_description", "=", "False", "python_spec_file", "=", "[", "]", "for", "line", "in", "spec_file", ":", "if", "line", ".", "startswith", "(", "\"Summary: \"", ")", ":", "summary", "=", "line", "elif", "line", ".", "startswith", "(", "\"BuildRequires: \"", ")", ":", "line", "=", "\"BuildRequires: {0:s}-setuptools\"", ".", "format", "(", "python_package", ")", "elif", "line", ".", "startswith", "(", "\"Requires: \"", ")", ":", "if", "python_package", "==", "\"python3\"", ":", "line", "=", "line", ".", "replace", "(", "\"python\"", ",", "\"python3\"", ")", "elif", "line", ".", "startswith", "(", "\"%description\"", ")", ":", "in_description", "=", "True", "elif", "line", ".", "startswith", "(", "\"%files\"", ")", ":", "line", "=", "\"%files -f INSTALLED_FILES -n {0:s}-%{{name}}\"", ".", "format", "(", "python_package", ")", "elif", "line", ".", "startswith", "(", "\"%prep\"", ")", ":", "in_description", "=", "False", "python_spec_file", ".", "append", "(", "\"%package -n {0:s}-%{{name}}\"", ".", "format", "(", "python_package", ")", ")", "python_spec_file", ".", "append", "(", "\"{0:s}\"", ".", "format", "(", "summary", ")", ")", "python_spec_file", ".", "append", "(", "\"\"", ")", "python_spec_file", ".", "append", "(", "\"%description -n {0:s}-%{{name}}\"", ".", "format", "(", "python_package", ")", ")", "python_spec_file", ".", "extend", "(", "description", ")", "elif", "in_description", ":", "# Ignore leading white lines in the description.", "if", "not", "description", "and", "not", "line", ":", "continue", "description", ".", "append", "(", "line", ")", "python_spec_file", ".", "append", "(", "line", ")", "return", "python_spec_file" ]
Generates the text of an RPM spec file. Returns: A list of strings containing the lines of text.
[ "Generates", "the", "text", "of", "an", "RPM", "spec", "file", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/setup.py#L45-L105
google/dotty
efilter/scope.py
ScopeStack.resolve
def resolve(self, name): """Call IStructured.resolve across all scopes and return first hit.""" for scope in reversed(self.scopes): try: return structured.resolve(scope, name) except (KeyError, AttributeError): continue raise AttributeError(name)
python
def resolve(self, name): """Call IStructured.resolve across all scopes and return first hit.""" for scope in reversed(self.scopes): try: return structured.resolve(scope, name) except (KeyError, AttributeError): continue raise AttributeError(name)
[ "def", "resolve", "(", "self", ",", "name", ")", ":", "for", "scope", "in", "reversed", "(", "self", ".", "scopes", ")", ":", "try", ":", "return", "structured", ".", "resolve", "(", "scope", ",", "name", ")", "except", "(", "KeyError", ",", "AttributeError", ")", ":", "continue", "raise", "AttributeError", "(", "name", ")" ]
Call IStructured.resolve across all scopes and return first hit.
[ "Call", "IStructured", ".", "resolve", "across", "all", "scopes", "and", "return", "first", "hit", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L73-L81
google/dotty
efilter/scope.py
ScopeStack.getmembers
def getmembers(self): """Gets members (vars) from all scopes, using both runtime and static. This method will attempt both static and runtime getmembers. This is the recommended way of getting available members. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'. """ names = set() for scope in self.scopes: if isinstance(scope, type): names.update(structured.getmembers_static(scope)) else: names.update(structured.getmembers_runtime(scope)) return names
python
def getmembers(self): """Gets members (vars) from all scopes, using both runtime and static. This method will attempt both static and runtime getmembers. This is the recommended way of getting available members. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'. """ names = set() for scope in self.scopes: if isinstance(scope, type): names.update(structured.getmembers_static(scope)) else: names.update(structured.getmembers_runtime(scope)) return names
[ "def", "getmembers", "(", "self", ")", ":", "names", "=", "set", "(", ")", "for", "scope", "in", "self", ".", "scopes", ":", "if", "isinstance", "(", "scope", ",", "type", ")", ":", "names", ".", "update", "(", "structured", ".", "getmembers_static", "(", "scope", ")", ")", "else", ":", "names", ".", "update", "(", "structured", ".", "getmembers_runtime", "(", "scope", ")", ")", "return", "names" ]
Gets members (vars) from all scopes, using both runtime and static. This method will attempt both static and runtime getmembers. This is the recommended way of getting available members. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'.
[ "Gets", "members", "(", "vars", ")", "from", "all", "scopes", "using", "both", "runtime", "and", "static", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L83-L102
google/dotty
efilter/scope.py
ScopeStack.getmembers_runtime
def getmembers_runtime(self): """Gets members (vars) from all scopes using ONLY runtime information. You most likely want to use ScopeStack.getmembers instead. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'. """ names = set() for scope in self.scopes: names.update(structured.getmembers_runtime(scope)) return names
python
def getmembers_runtime(self): """Gets members (vars) from all scopes using ONLY runtime information. You most likely want to use ScopeStack.getmembers instead. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'. """ names = set() for scope in self.scopes: names.update(structured.getmembers_runtime(scope)) return names
[ "def", "getmembers_runtime", "(", "self", ")", ":", "names", "=", "set", "(", ")", "for", "scope", "in", "self", ".", "scopes", ":", "names", ".", "update", "(", "structured", ".", "getmembers_runtime", "(", "scope", ")", ")", "return", "names" ]
Gets members (vars) from all scopes using ONLY runtime information. You most likely want to use ScopeStack.getmembers instead. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'.
[ "Gets", "members", "(", "vars", ")", "from", "all", "scopes", "using", "ONLY", "runtime", "information", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L104-L119
google/dotty
efilter/scope.py
ScopeStack.getmembers_static
def getmembers_static(cls): """Gets members (vars) from all scopes using ONLY static information. You most likely want to use ScopeStack.getmembers instead. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'. """ names = set() for scope in cls.scopes: names.update(structured.getmembers_static(scope)) return names
python
def getmembers_static(cls): """Gets members (vars) from all scopes using ONLY static information. You most likely want to use ScopeStack.getmembers instead. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'. """ names = set() for scope in cls.scopes: names.update(structured.getmembers_static(scope)) return names
[ "def", "getmembers_static", "(", "cls", ")", ":", "names", "=", "set", "(", ")", "for", "scope", "in", "cls", ".", "scopes", ":", "names", ".", "update", "(", "structured", ".", "getmembers_static", "(", "scope", ")", ")", "return", "names" ]
Gets members (vars) from all scopes using ONLY static information. You most likely want to use ScopeStack.getmembers instead. Returns: Set of available vars. Raises: NotImplementedError if any scope fails to implement 'getmembers'.
[ "Gets", "members", "(", "vars", ")", "from", "all", "scopes", "using", "ONLY", "static", "information", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L122-L137
google/dotty
efilter/scope.py
ScopeStack.reflect
def reflect(self, name): """Reflect 'name' starting with local scope all the way up to global. This method will attempt both static and runtime reflection. This is the recommended way of using reflection. Returns: Type of 'name', or protocol.AnyType. Caveat: The type of 'name' does not necessarily have to be an instance of Python's type - it depends on what the host application returns through the reflection API. For example, Rekall uses objects generated at runtime to simulate a native (C/C++) type system. """ # Return whatever the most local scope defines this as, or bubble all # the way to the top. result = None for scope in reversed(self.scopes): try: if isinstance(scope, type): result = structured.reflect_static_member(scope, name) else: result = structured.reflect_runtime_member(scope, name) if result is not None: return result except (NotImplementedError, KeyError, AttributeError): continue return protocol.AnyType
python
def reflect(self, name): """Reflect 'name' starting with local scope all the way up to global. This method will attempt both static and runtime reflection. This is the recommended way of using reflection. Returns: Type of 'name', or protocol.AnyType. Caveat: The type of 'name' does not necessarily have to be an instance of Python's type - it depends on what the host application returns through the reflection API. For example, Rekall uses objects generated at runtime to simulate a native (C/C++) type system. """ # Return whatever the most local scope defines this as, or bubble all # the way to the top. result = None for scope in reversed(self.scopes): try: if isinstance(scope, type): result = structured.reflect_static_member(scope, name) else: result = structured.reflect_runtime_member(scope, name) if result is not None: return result except (NotImplementedError, KeyError, AttributeError): continue return protocol.AnyType
[ "def", "reflect", "(", "self", ",", "name", ")", ":", "# Return whatever the most local scope defines this as, or bubble all", "# the way to the top.", "result", "=", "None", "for", "scope", "in", "reversed", "(", "self", ".", "scopes", ")", ":", "try", ":", "if", "isinstance", "(", "scope", ",", "type", ")", ":", "result", "=", "structured", ".", "reflect_static_member", "(", "scope", ",", "name", ")", "else", ":", "result", "=", "structured", ".", "reflect_runtime_member", "(", "scope", ",", "name", ")", "if", "result", "is", "not", "None", ":", "return", "result", "except", "(", "NotImplementedError", ",", "KeyError", ",", "AttributeError", ")", ":", "continue", "return", "protocol", ".", "AnyType" ]
Reflect 'name' starting with local scope all the way up to global. This method will attempt both static and runtime reflection. This is the recommended way of using reflection. Returns: Type of 'name', or protocol.AnyType. Caveat: The type of 'name' does not necessarily have to be an instance of Python's type - it depends on what the host application returns through the reflection API. For example, Rekall uses objects generated at runtime to simulate a native (C/C++) type system.
[ "Reflect", "name", "starting", "with", "local", "scope", "all", "the", "way", "up", "to", "global", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L139-L170
google/dotty
efilter/scope.py
ScopeStack.reflect_runtime_member
def reflect_runtime_member(self, name): """Reflect 'name' using ONLY runtime reflection. You most likely want to use ScopeStack.reflect instead. Returns: Type of 'name', or protocol.AnyType. """ for scope in reversed(self.scopes): try: return structured.reflect_runtime_member(scope, name) except (NotImplementedError, KeyError, AttributeError): continue return protocol.AnyType
python
def reflect_runtime_member(self, name): """Reflect 'name' using ONLY runtime reflection. You most likely want to use ScopeStack.reflect instead. Returns: Type of 'name', or protocol.AnyType. """ for scope in reversed(self.scopes): try: return structured.reflect_runtime_member(scope, name) except (NotImplementedError, KeyError, AttributeError): continue return protocol.AnyType
[ "def", "reflect_runtime_member", "(", "self", ",", "name", ")", ":", "for", "scope", "in", "reversed", "(", "self", ".", "scopes", ")", ":", "try", ":", "return", "structured", ".", "reflect_runtime_member", "(", "scope", ",", "name", ")", "except", "(", "NotImplementedError", ",", "KeyError", ",", "AttributeError", ")", ":", "continue", "return", "protocol", ".", "AnyType" ]
Reflect 'name' using ONLY runtime reflection. You most likely want to use ScopeStack.reflect instead. Returns: Type of 'name', or protocol.AnyType.
[ "Reflect", "name", "using", "ONLY", "runtime", "reflection", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L172-L186
google/dotty
efilter/scope.py
ScopeStack.reflect_static_member
def reflect_static_member(cls, name): """Reflect 'name' using ONLY static reflection. You most likely want to use ScopeStack.reflect instead. Returns: Type of 'name', or protocol.AnyType. """ for scope in reversed(cls.scopes): try: return structured.reflect_static_member(scope, name) except (NotImplementedError, KeyError, AttributeError): continue return protocol.AnyType
python
def reflect_static_member(cls, name): """Reflect 'name' using ONLY static reflection. You most likely want to use ScopeStack.reflect instead. Returns: Type of 'name', or protocol.AnyType. """ for scope in reversed(cls.scopes): try: return structured.reflect_static_member(scope, name) except (NotImplementedError, KeyError, AttributeError): continue return protocol.AnyType
[ "def", "reflect_static_member", "(", "cls", ",", "name", ")", ":", "for", "scope", "in", "reversed", "(", "cls", ".", "scopes", ")", ":", "try", ":", "return", "structured", ".", "reflect_static_member", "(", "scope", ",", "name", ")", "except", "(", "NotImplementedError", ",", "KeyError", ",", "AttributeError", ")", ":", "continue", "return", "protocol", ".", "AnyType" ]
Reflect 'name' using ONLY static reflection. You most likely want to use ScopeStack.reflect instead. Returns: Type of 'name', or protocol.AnyType.
[ "Reflect", "name", "using", "ONLY", "static", "reflection", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/scope.py#L189-L203
radical-cybertools/radical.entk
src/radical/entk/utils/prof_utils.py
get_hostmap
def get_hostmap(profile): ''' We abuse the profile combination to also derive a pilot-host map, which will tell us on what exact host each pilot has been running. To do so, we check for the PMGR_ACTIVE advance event in agent_0.prof, and use the NTP sync info to associate a hostname. ''' # FIXME: This should be replaced by proper hostname logging # in `pilot.resource_details`. hostmap = dict() # map pilot IDs to host names for entry in profile: if entry[ru.EVENT] == 'hostname': hostmap[entry[ru.UID]] = entry[ru.MSG] return hostmap
python
def get_hostmap(profile): ''' We abuse the profile combination to also derive a pilot-host map, which will tell us on what exact host each pilot has been running. To do so, we check for the PMGR_ACTIVE advance event in agent_0.prof, and use the NTP sync info to associate a hostname. ''' # FIXME: This should be replaced by proper hostname logging # in `pilot.resource_details`. hostmap = dict() # map pilot IDs to host names for entry in profile: if entry[ru.EVENT] == 'hostname': hostmap[entry[ru.UID]] = entry[ru.MSG] return hostmap
[ "def", "get_hostmap", "(", "profile", ")", ":", "# FIXME: This should be replaced by proper hostname logging", "# in `pilot.resource_details`.", "hostmap", "=", "dict", "(", ")", "# map pilot IDs to host names", "for", "entry", "in", "profile", ":", "if", "entry", "[", "ru", ".", "EVENT", "]", "==", "'hostname'", ":", "hostmap", "[", "entry", "[", "ru", ".", "UID", "]", "]", "=", "entry", "[", "ru", ".", "MSG", "]", "return", "hostmap" ]
We abuse the profile combination to also derive a pilot-host map, which will tell us on what exact host each pilot has been running. To do so, we check for the PMGR_ACTIVE advance event in agent_0.prof, and use the NTP sync info to associate a hostname.
[ "We", "abuse", "the", "profile", "combination", "to", "also", "derive", "a", "pilot", "-", "host", "map", "which", "will", "tell", "us", "on", "what", "exact", "host", "each", "pilot", "has", "been", "running", ".", "To", "do", "so", "we", "check", "for", "the", "PMGR_ACTIVE", "advance", "event", "in", "agent_0", ".", "prof", "and", "use", "the", "NTP", "sync", "info", "to", "associate", "a", "hostname", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/utils/prof_utils.py#L16-L31
radical-cybertools/radical.entk
src/radical/entk/utils/prof_utils.py
get_hostmap_deprecated
def get_hostmap_deprecated(profiles): ''' This method mangles combine_profiles and get_hostmap, and is deprecated. At this point it only returns the hostmap ''' hostmap = dict() # map pilot IDs to host names for pname, prof in profiles.iteritems(): if not len(prof): continue if not prof[0][ru.MSG]: continue host, ip, _, _, _ = prof[0][ru.MSG].split(':') host_id = '%s:%s' % (host, ip) for row in prof: if 'agent_0.prof' in pname and \ row[ru.EVENT] == 'advance' and \ row[ru.STATE] == rps.PMGR_ACTIVE: hostmap[row[ru.UID]] = host_id break return hostmap
python
def get_hostmap_deprecated(profiles): ''' This method mangles combine_profiles and get_hostmap, and is deprecated. At this point it only returns the hostmap ''' hostmap = dict() # map pilot IDs to host names for pname, prof in profiles.iteritems(): if not len(prof): continue if not prof[0][ru.MSG]: continue host, ip, _, _, _ = prof[0][ru.MSG].split(':') host_id = '%s:%s' % (host, ip) for row in prof: if 'agent_0.prof' in pname and \ row[ru.EVENT] == 'advance' and \ row[ru.STATE] == rps.PMGR_ACTIVE: hostmap[row[ru.UID]] = host_id break return hostmap
[ "def", "get_hostmap_deprecated", "(", "profiles", ")", ":", "hostmap", "=", "dict", "(", ")", "# map pilot IDs to host names", "for", "pname", ",", "prof", "in", "profiles", ".", "iteritems", "(", ")", ":", "if", "not", "len", "(", "prof", ")", ":", "continue", "if", "not", "prof", "[", "0", "]", "[", "ru", ".", "MSG", "]", ":", "continue", "host", ",", "ip", ",", "_", ",", "_", ",", "_", "=", "prof", "[", "0", "]", "[", "ru", ".", "MSG", "]", ".", "split", "(", "':'", ")", "host_id", "=", "'%s:%s'", "%", "(", "host", ",", "ip", ")", "for", "row", "in", "prof", ":", "if", "'agent_0.prof'", "in", "pname", "and", "row", "[", "ru", ".", "EVENT", "]", "==", "'advance'", "and", "row", "[", "ru", ".", "STATE", "]", "==", "rps", ".", "PMGR_ACTIVE", ":", "hostmap", "[", "row", "[", "ru", ".", "UID", "]", "]", "=", "host_id", "break", "return", "hostmap" ]
This method mangles combine_profiles and get_hostmap, and is deprecated. At this point it only returns the hostmap
[ "This", "method", "mangles", "combine_profiles", "and", "get_hostmap", "and", "is", "deprecated", ".", "At", "this", "point", "it", "only", "returns", "the", "hostmap" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/utils/prof_utils.py#L34-L60
radical-cybertools/radical.entk
src/radical/entk/appman/appmanager.py
AppManager.run
def run(self): """ **Purpose**: Run the application manager. Once the workflow and resource manager have been assigned. Invoking this method will start the setting up the communication infrastructure, submitting a resource request and then submission of all the tasks. """ try: # Set None objects local to each run self._wfp = None self._sync_thread = None self._terminate_sync = Event() self._resubmit_failed = False self._cur_attempt = 1 if not self._workflow: self._logger.error('No workflow assigned currently, please check your script') raise MissingError(obj=self._uid, missing_attribute='workflow') if not self._resource_manager: self._logger.error('No resource manager assigned currently, please create and add a valid resource manager') raise MissingError(obj=self._uid, missing_attribute='resource_manager') self._prof.prof('amgr run started', uid=self._uid) # Setup rabbitmq stuff if not self._mqs_setup: self._report.info('Setting up RabbitMQ system') setup = self._setup_mqs() if not setup: self._logger.error('RabbitMQ system not available') raise EnTKError("RabbitMQ setup failed") self._mqs_setup = True self._report.ok('>>ok\n') # Create WFProcessor object self._prof.prof('creating wfp obj', uid=self._uid) self._wfp = WFprocessor(sid=self._sid, workflow=self._workflow, pending_queue=self._pending_queue, completed_queue=self._completed_queue, mq_hostname=self._mq_hostname, port=self._port, resubmit_failed=self._resubmit_failed) self._wfp._initialize_workflow() self._workflow = self._wfp.workflow # Submit resource request if not resource allocation done till now or # resubmit a new one if the old one has completed if self._resource_manager: res_alloc_state = self._resource_manager.get_resource_allocation_state() if (not res_alloc_state) or (res_alloc_state in self._resource_manager.get_completed_states()): self._logger.info('Starting resource request submission') self._prof.prof('init rreq submission', uid=self._uid) self._resource_manager._submit_resource_request() res_alloc_state = self._resource_manager.get_resource_allocation_state() if res_alloc_state in self._resource_manager.get_completed_states(): raise EnTKError(msg="Cannot proceed. Resource allocation ended up in %s"%res_alloc_state) else: self._logger.exception('Cannot run without resource manager, please create and assign a resource manager') raise EnTKError(text='Missing resource manager') # Start synchronizer thread if not self._sync_thread: self._logger.info('Starting synchronizer thread') self._sync_thread = Thread(target=self._synchronizer, name='synchronizer-thread') self._prof.prof('starting synchronizer thread', uid=self._uid) self._sync_thread.start() # Start WFprocessor self._logger.info('Starting WFProcessor process from AppManager') self._wfp.start_processor() self._report.ok('All components created\n') # Create tmgr object only if it does not already exist if self._rts == 'radical.pilot': from radical.entk.execman.rp import TaskManager elif self._rts == 'mock': from radical.entk.execman.mock import TaskManager if not self._task_manager: self._prof.prof('creating tmgr obj', uid=self._uid) self._task_manager = TaskManager(sid=self._sid, pending_queue=self._pending_queue, completed_queue=self._completed_queue, mq_hostname=self._mq_hostname, rmgr=self._resource_manager, port=self._port ) self._logger.info('Starting task manager process from AppManager') self._task_manager.start_manager() self._task_manager.start_heartbeat() active_pipe_count = len(self._workflow) finished_pipe_uids = [] # We wait till all pipelines of the workflow are marked # complete while ((active_pipe_count > 0) and (self._wfp.workflow_incomplete()) and (self._resource_manager.get_resource_allocation_state() not in self._resource_manager.get_completed_states())): if active_pipe_count > 0: for pipe in self._workflow: with pipe.lock: if (pipe.completed) and (pipe.uid not in finished_pipe_uids): self._logger.info('Pipe %s completed' % pipe.uid) finished_pipe_uids.append(pipe.uid) active_pipe_count -= 1 self._logger.info('Active pipes: %s' % active_pipe_count) if (not self._sync_thread.is_alive()) and (self._cur_attempt <= self._reattempts): self._sync_thread = Thread(target=self._synchronizer, name='synchronizer-thread') self._logger.info('Restarting synchronizer thread') self._prof.prof('restarting synchronizer', uid=self._uid) self._sync_thread.start() self._cur_attempt += 1 if (not self._wfp.check_processor()) and (self._cur_attempt <= self._reattempts): """ If WFP dies, both child threads are also cleaned out. We simply recreate the wfp object with a copy of the workflow in the appmanager and start the processor. """ self._prof.prof('recreating wfp obj', uid=self._uid) self._wfp = WFProcessor( sid=self._sid, workflow=self._workflow, pending_queue=self._pending_queue, completed_queue=self._completed_queue, mq_hostname=self._mq_hostname, port=self._port, resubmit_failed=self._resubmit_failed) self._logger.info('Restarting WFProcessor process from AppManager') self._wfp.start_processor() self._cur_attempt += 1 if (not self._task_manager.check_heartbeat()) and (self._cur_attempt <= self._reattempts): """ If the tmgr process or heartbeat dies, we simply start a new process using the start_manager method. We do not need to create a new instance of the TaskManager object itself. We stop and start a new instance of the heartbeat thread as well. """ self._prof.prof('restarting tmgr process and heartbeat', uid=self._uid) self._logger.info('Terminating heartbeat thread') self._task_manager.terminate_heartbeat() self._logger.info('Terminating tmgr process') self._task_manager.terminate_manager() self._logger.info('Restarting task manager process') self._task_manager.start_manager() self._logger.info('Restarting heartbeat thread') self._task_manager.start_heartbeat() self._cur_attempt += 1 self._prof.prof('start termination', uid=self._uid) # Terminate threads in following order: wfp, helper, synchronizer self._logger.info('Terminating WFprocessor') self._wfp.terminate_processor() self._logger.info('Terminating synchronizer thread') self._terminate_sync.set() self._sync_thread.join() self._logger.info('Synchronizer thread terminated') if self._autoterminate: self.resource_terminate() if self._write_workflow: write_workflow(self._workflow, self._sid) self._prof.prof('termination done', uid=self._uid) except KeyboardInterrupt: self._prof.prof('start termination', uid=self._uid) self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel enqueuer thread gracefully...') # Terminate threads in following order: wfp, helper, synchronizer if self._wfp: self._logger.info('Terminating WFprocessor') self._wfp.terminate_processor() if self._task_manager: self._logger.info('Terminating task manager process') self._task_manager.terminate_manager() self._task_manager.terminate_heartbeat() if self._sync_thread: self._logger.info('Terminating synchronizer thread') self._terminate_sync.set() self._sync_thread.join() self._logger.info('Synchronizer thread terminated') if self._resource_manager: self._resource_manager._terminate_resource_request() self._prof.prof('termination done', uid=self._uid) raise KeyboardInterrupt except Exception, ex: self._prof.prof('start termination', uid=self._uid) self._logger.exception('Error in AppManager: %s' % ex) # Terminate threads in following order: wfp, helper, synchronizer if self._wfp: self._logger.info('Terminating WFprocessor') self._wfp.terminate_processor() if self._task_manager: self._logger.info('Terminating task manager process') self._task_manager.terminate_manager() self._task_manager.terminate_heartbeat() if self._sync_thread: self._logger.info('Terminating synchronizer thread') self._terminate_sync.set() self._sync_thread.join() self._logger.info('Synchronizer thread terminated') if self._resource_manager: self._resource_manager._terminate_resource_request() self._prof.prof('termination done', uid=self._uid) raise
python
def run(self): """ **Purpose**: Run the application manager. Once the workflow and resource manager have been assigned. Invoking this method will start the setting up the communication infrastructure, submitting a resource request and then submission of all the tasks. """ try: # Set None objects local to each run self._wfp = None self._sync_thread = None self._terminate_sync = Event() self._resubmit_failed = False self._cur_attempt = 1 if not self._workflow: self._logger.error('No workflow assigned currently, please check your script') raise MissingError(obj=self._uid, missing_attribute='workflow') if not self._resource_manager: self._logger.error('No resource manager assigned currently, please create and add a valid resource manager') raise MissingError(obj=self._uid, missing_attribute='resource_manager') self._prof.prof('amgr run started', uid=self._uid) # Setup rabbitmq stuff if not self._mqs_setup: self._report.info('Setting up RabbitMQ system') setup = self._setup_mqs() if not setup: self._logger.error('RabbitMQ system not available') raise EnTKError("RabbitMQ setup failed") self._mqs_setup = True self._report.ok('>>ok\n') # Create WFProcessor object self._prof.prof('creating wfp obj', uid=self._uid) self._wfp = WFprocessor(sid=self._sid, workflow=self._workflow, pending_queue=self._pending_queue, completed_queue=self._completed_queue, mq_hostname=self._mq_hostname, port=self._port, resubmit_failed=self._resubmit_failed) self._wfp._initialize_workflow() self._workflow = self._wfp.workflow # Submit resource request if not resource allocation done till now or # resubmit a new one if the old one has completed if self._resource_manager: res_alloc_state = self._resource_manager.get_resource_allocation_state() if (not res_alloc_state) or (res_alloc_state in self._resource_manager.get_completed_states()): self._logger.info('Starting resource request submission') self._prof.prof('init rreq submission', uid=self._uid) self._resource_manager._submit_resource_request() res_alloc_state = self._resource_manager.get_resource_allocation_state() if res_alloc_state in self._resource_manager.get_completed_states(): raise EnTKError(msg="Cannot proceed. Resource allocation ended up in %s"%res_alloc_state) else: self._logger.exception('Cannot run without resource manager, please create and assign a resource manager') raise EnTKError(text='Missing resource manager') # Start synchronizer thread if not self._sync_thread: self._logger.info('Starting synchronizer thread') self._sync_thread = Thread(target=self._synchronizer, name='synchronizer-thread') self._prof.prof('starting synchronizer thread', uid=self._uid) self._sync_thread.start() # Start WFprocessor self._logger.info('Starting WFProcessor process from AppManager') self._wfp.start_processor() self._report.ok('All components created\n') # Create tmgr object only if it does not already exist if self._rts == 'radical.pilot': from radical.entk.execman.rp import TaskManager elif self._rts == 'mock': from radical.entk.execman.mock import TaskManager if not self._task_manager: self._prof.prof('creating tmgr obj', uid=self._uid) self._task_manager = TaskManager(sid=self._sid, pending_queue=self._pending_queue, completed_queue=self._completed_queue, mq_hostname=self._mq_hostname, rmgr=self._resource_manager, port=self._port ) self._logger.info('Starting task manager process from AppManager') self._task_manager.start_manager() self._task_manager.start_heartbeat() active_pipe_count = len(self._workflow) finished_pipe_uids = [] # We wait till all pipelines of the workflow are marked # complete while ((active_pipe_count > 0) and (self._wfp.workflow_incomplete()) and (self._resource_manager.get_resource_allocation_state() not in self._resource_manager.get_completed_states())): if active_pipe_count > 0: for pipe in self._workflow: with pipe.lock: if (pipe.completed) and (pipe.uid not in finished_pipe_uids): self._logger.info('Pipe %s completed' % pipe.uid) finished_pipe_uids.append(pipe.uid) active_pipe_count -= 1 self._logger.info('Active pipes: %s' % active_pipe_count) if (not self._sync_thread.is_alive()) and (self._cur_attempt <= self._reattempts): self._sync_thread = Thread(target=self._synchronizer, name='synchronizer-thread') self._logger.info('Restarting synchronizer thread') self._prof.prof('restarting synchronizer', uid=self._uid) self._sync_thread.start() self._cur_attempt += 1 if (not self._wfp.check_processor()) and (self._cur_attempt <= self._reattempts): """ If WFP dies, both child threads are also cleaned out. We simply recreate the wfp object with a copy of the workflow in the appmanager and start the processor. """ self._prof.prof('recreating wfp obj', uid=self._uid) self._wfp = WFProcessor( sid=self._sid, workflow=self._workflow, pending_queue=self._pending_queue, completed_queue=self._completed_queue, mq_hostname=self._mq_hostname, port=self._port, resubmit_failed=self._resubmit_failed) self._logger.info('Restarting WFProcessor process from AppManager') self._wfp.start_processor() self._cur_attempt += 1 if (not self._task_manager.check_heartbeat()) and (self._cur_attempt <= self._reattempts): """ If the tmgr process or heartbeat dies, we simply start a new process using the start_manager method. We do not need to create a new instance of the TaskManager object itself. We stop and start a new instance of the heartbeat thread as well. """ self._prof.prof('restarting tmgr process and heartbeat', uid=self._uid) self._logger.info('Terminating heartbeat thread') self._task_manager.terminate_heartbeat() self._logger.info('Terminating tmgr process') self._task_manager.terminate_manager() self._logger.info('Restarting task manager process') self._task_manager.start_manager() self._logger.info('Restarting heartbeat thread') self._task_manager.start_heartbeat() self._cur_attempt += 1 self._prof.prof('start termination', uid=self._uid) # Terminate threads in following order: wfp, helper, synchronizer self._logger.info('Terminating WFprocessor') self._wfp.terminate_processor() self._logger.info('Terminating synchronizer thread') self._terminate_sync.set() self._sync_thread.join() self._logger.info('Synchronizer thread terminated') if self._autoterminate: self.resource_terminate() if self._write_workflow: write_workflow(self._workflow, self._sid) self._prof.prof('termination done', uid=self._uid) except KeyboardInterrupt: self._prof.prof('start termination', uid=self._uid) self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel enqueuer thread gracefully...') # Terminate threads in following order: wfp, helper, synchronizer if self._wfp: self._logger.info('Terminating WFprocessor') self._wfp.terminate_processor() if self._task_manager: self._logger.info('Terminating task manager process') self._task_manager.terminate_manager() self._task_manager.terminate_heartbeat() if self._sync_thread: self._logger.info('Terminating synchronizer thread') self._terminate_sync.set() self._sync_thread.join() self._logger.info('Synchronizer thread terminated') if self._resource_manager: self._resource_manager._terminate_resource_request() self._prof.prof('termination done', uid=self._uid) raise KeyboardInterrupt except Exception, ex: self._prof.prof('start termination', uid=self._uid) self._logger.exception('Error in AppManager: %s' % ex) # Terminate threads in following order: wfp, helper, synchronizer if self._wfp: self._logger.info('Terminating WFprocessor') self._wfp.terminate_processor() if self._task_manager: self._logger.info('Terminating task manager process') self._task_manager.terminate_manager() self._task_manager.terminate_heartbeat() if self._sync_thread: self._logger.info('Terminating synchronizer thread') self._terminate_sync.set() self._sync_thread.join() self._logger.info('Synchronizer thread terminated') if self._resource_manager: self._resource_manager._terminate_resource_request() self._prof.prof('termination done', uid=self._uid) raise
[ "def", "run", "(", "self", ")", ":", "try", ":", "# Set None objects local to each run", "self", ".", "_wfp", "=", "None", "self", ".", "_sync_thread", "=", "None", "self", ".", "_terminate_sync", "=", "Event", "(", ")", "self", ".", "_resubmit_failed", "=", "False", "self", ".", "_cur_attempt", "=", "1", "if", "not", "self", ".", "_workflow", ":", "self", ".", "_logger", ".", "error", "(", "'No workflow assigned currently, please check your script'", ")", "raise", "MissingError", "(", "obj", "=", "self", ".", "_uid", ",", "missing_attribute", "=", "'workflow'", ")", "if", "not", "self", ".", "_resource_manager", ":", "self", ".", "_logger", ".", "error", "(", "'No resource manager assigned currently, please create and add a valid resource manager'", ")", "raise", "MissingError", "(", "obj", "=", "self", ".", "_uid", ",", "missing_attribute", "=", "'resource_manager'", ")", "self", ".", "_prof", ".", "prof", "(", "'amgr run started'", ",", "uid", "=", "self", ".", "_uid", ")", "# Setup rabbitmq stuff", "if", "not", "self", ".", "_mqs_setup", ":", "self", ".", "_report", ".", "info", "(", "'Setting up RabbitMQ system'", ")", "setup", "=", "self", ".", "_setup_mqs", "(", ")", "if", "not", "setup", ":", "self", ".", "_logger", ".", "error", "(", "'RabbitMQ system not available'", ")", "raise", "EnTKError", "(", "\"RabbitMQ setup failed\"", ")", "self", ".", "_mqs_setup", "=", "True", "self", ".", "_report", ".", "ok", "(", "'>>ok\\n'", ")", "# Create WFProcessor object", "self", ".", "_prof", ".", "prof", "(", "'creating wfp obj'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_wfp", "=", "WFprocessor", "(", "sid", "=", "self", ".", "_sid", ",", "workflow", "=", "self", ".", "_workflow", ",", "pending_queue", "=", "self", ".", "_pending_queue", ",", "completed_queue", "=", "self", ".", "_completed_queue", ",", "mq_hostname", "=", "self", ".", "_mq_hostname", ",", "port", "=", "self", ".", "_port", ",", "resubmit_failed", "=", "self", ".", "_resubmit_failed", ")", "self", ".", "_wfp", ".", "_initialize_workflow", "(", ")", "self", ".", "_workflow", "=", "self", ".", "_wfp", ".", "workflow", "# Submit resource request if not resource allocation done till now or", "# resubmit a new one if the old one has completed", "if", "self", ".", "_resource_manager", ":", "res_alloc_state", "=", "self", ".", "_resource_manager", ".", "get_resource_allocation_state", "(", ")", "if", "(", "not", "res_alloc_state", ")", "or", "(", "res_alloc_state", "in", "self", ".", "_resource_manager", ".", "get_completed_states", "(", ")", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Starting resource request submission'", ")", "self", ".", "_prof", ".", "prof", "(", "'init rreq submission'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_resource_manager", ".", "_submit_resource_request", "(", ")", "res_alloc_state", "=", "self", ".", "_resource_manager", ".", "get_resource_allocation_state", "(", ")", "if", "res_alloc_state", "in", "self", ".", "_resource_manager", ".", "get_completed_states", "(", ")", ":", "raise", "EnTKError", "(", "msg", "=", "\"Cannot proceed. Resource allocation ended up in %s\"", "%", "res_alloc_state", ")", "else", ":", "self", ".", "_logger", ".", "exception", "(", "'Cannot run without resource manager, please create and assign a resource manager'", ")", "raise", "EnTKError", "(", "text", "=", "'Missing resource manager'", ")", "# Start synchronizer thread", "if", "not", "self", ".", "_sync_thread", ":", "self", ".", "_logger", ".", "info", "(", "'Starting synchronizer thread'", ")", "self", ".", "_sync_thread", "=", "Thread", "(", "target", "=", "self", ".", "_synchronizer", ",", "name", "=", "'synchronizer-thread'", ")", "self", ".", "_prof", ".", "prof", "(", "'starting synchronizer thread'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_sync_thread", ".", "start", "(", ")", "# Start WFprocessor", "self", ".", "_logger", ".", "info", "(", "'Starting WFProcessor process from AppManager'", ")", "self", ".", "_wfp", ".", "start_processor", "(", ")", "self", ".", "_report", ".", "ok", "(", "'All components created\\n'", ")", "# Create tmgr object only if it does not already exist", "if", "self", ".", "_rts", "==", "'radical.pilot'", ":", "from", "radical", ".", "entk", ".", "execman", ".", "rp", "import", "TaskManager", "elif", "self", ".", "_rts", "==", "'mock'", ":", "from", "radical", ".", "entk", ".", "execman", ".", "mock", "import", "TaskManager", "if", "not", "self", ".", "_task_manager", ":", "self", ".", "_prof", ".", "prof", "(", "'creating tmgr obj'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_task_manager", "=", "TaskManager", "(", "sid", "=", "self", ".", "_sid", ",", "pending_queue", "=", "self", ".", "_pending_queue", ",", "completed_queue", "=", "self", ".", "_completed_queue", ",", "mq_hostname", "=", "self", ".", "_mq_hostname", ",", "rmgr", "=", "self", ".", "_resource_manager", ",", "port", "=", "self", ".", "_port", ")", "self", ".", "_logger", ".", "info", "(", "'Starting task manager process from AppManager'", ")", "self", ".", "_task_manager", ".", "start_manager", "(", ")", "self", ".", "_task_manager", ".", "start_heartbeat", "(", ")", "active_pipe_count", "=", "len", "(", "self", ".", "_workflow", ")", "finished_pipe_uids", "=", "[", "]", "# We wait till all pipelines of the workflow are marked", "# complete", "while", "(", "(", "active_pipe_count", ">", "0", ")", "and", "(", "self", ".", "_wfp", ".", "workflow_incomplete", "(", ")", ")", "and", "(", "self", ".", "_resource_manager", ".", "get_resource_allocation_state", "(", ")", "not", "in", "self", ".", "_resource_manager", ".", "get_completed_states", "(", ")", ")", ")", ":", "if", "active_pipe_count", ">", "0", ":", "for", "pipe", "in", "self", ".", "_workflow", ":", "with", "pipe", ".", "lock", ":", "if", "(", "pipe", ".", "completed", ")", "and", "(", "pipe", ".", "uid", "not", "in", "finished_pipe_uids", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Pipe %s completed'", "%", "pipe", ".", "uid", ")", "finished_pipe_uids", ".", "append", "(", "pipe", ".", "uid", ")", "active_pipe_count", "-=", "1", "self", ".", "_logger", ".", "info", "(", "'Active pipes: %s'", "%", "active_pipe_count", ")", "if", "(", "not", "self", ".", "_sync_thread", ".", "is_alive", "(", ")", ")", "and", "(", "self", ".", "_cur_attempt", "<=", "self", ".", "_reattempts", ")", ":", "self", ".", "_sync_thread", "=", "Thread", "(", "target", "=", "self", ".", "_synchronizer", ",", "name", "=", "'synchronizer-thread'", ")", "self", ".", "_logger", ".", "info", "(", "'Restarting synchronizer thread'", ")", "self", ".", "_prof", ".", "prof", "(", "'restarting synchronizer'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_sync_thread", ".", "start", "(", ")", "self", ".", "_cur_attempt", "+=", "1", "if", "(", "not", "self", ".", "_wfp", ".", "check_processor", "(", ")", ")", "and", "(", "self", ".", "_cur_attempt", "<=", "self", ".", "_reattempts", ")", ":", "\"\"\"\n If WFP dies, both child threads are also cleaned out.\n We simply recreate the wfp object with a copy of the workflow\n in the appmanager and start the processor.\n \"\"\"", "self", ".", "_prof", ".", "prof", "(", "'recreating wfp obj'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_wfp", "=", "WFProcessor", "(", "sid", "=", "self", ".", "_sid", ",", "workflow", "=", "self", ".", "_workflow", ",", "pending_queue", "=", "self", ".", "_pending_queue", ",", "completed_queue", "=", "self", ".", "_completed_queue", ",", "mq_hostname", "=", "self", ".", "_mq_hostname", ",", "port", "=", "self", ".", "_port", ",", "resubmit_failed", "=", "self", ".", "_resubmit_failed", ")", "self", ".", "_logger", ".", "info", "(", "'Restarting WFProcessor process from AppManager'", ")", "self", ".", "_wfp", ".", "start_processor", "(", ")", "self", ".", "_cur_attempt", "+=", "1", "if", "(", "not", "self", ".", "_task_manager", ".", "check_heartbeat", "(", ")", ")", "and", "(", "self", ".", "_cur_attempt", "<=", "self", ".", "_reattempts", ")", ":", "\"\"\"\n If the tmgr process or heartbeat dies, we simply start a\n new process using the start_manager method. We do not\n need to create a new instance of the TaskManager object\n itself. We stop and start a new instance of the\n heartbeat thread as well.\n \"\"\"", "self", ".", "_prof", ".", "prof", "(", "'restarting tmgr process and heartbeat'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'Terminating heartbeat thread'", ")", "self", ".", "_task_manager", ".", "terminate_heartbeat", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Terminating tmgr process'", ")", "self", ".", "_task_manager", ".", "terminate_manager", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Restarting task manager process'", ")", "self", ".", "_task_manager", ".", "start_manager", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Restarting heartbeat thread'", ")", "self", ".", "_task_manager", ".", "start_heartbeat", "(", ")", "self", ".", "_cur_attempt", "+=", "1", "self", ".", "_prof", ".", "prof", "(", "'start termination'", ",", "uid", "=", "self", ".", "_uid", ")", "# Terminate threads in following order: wfp, helper, synchronizer", "self", ".", "_logger", ".", "info", "(", "'Terminating WFprocessor'", ")", "self", ".", "_wfp", ".", "terminate_processor", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Terminating synchronizer thread'", ")", "self", ".", "_terminate_sync", ".", "set", "(", ")", "self", ".", "_sync_thread", ".", "join", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Synchronizer thread terminated'", ")", "if", "self", ".", "_autoterminate", ":", "self", ".", "resource_terminate", "(", ")", "if", "self", ".", "_write_workflow", ":", "write_workflow", "(", "self", ".", "_workflow", ",", "self", ".", "_sid", ")", "self", ".", "_prof", ".", "prof", "(", "'termination done'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "KeyboardInterrupt", ":", "self", ".", "_prof", ".", "prof", "(", "'start termination'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to cancel enqueuer thread gracefully...'", ")", "# Terminate threads in following order: wfp, helper, synchronizer", "if", "self", ".", "_wfp", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating WFprocessor'", ")", "self", ".", "_wfp", ".", "terminate_processor", "(", ")", "if", "self", ".", "_task_manager", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating task manager process'", ")", "self", ".", "_task_manager", ".", "terminate_manager", "(", ")", "self", ".", "_task_manager", ".", "terminate_heartbeat", "(", ")", "if", "self", ".", "_sync_thread", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating synchronizer thread'", ")", "self", ".", "_terminate_sync", ".", "set", "(", ")", "self", ".", "_sync_thread", ".", "join", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Synchronizer thread terminated'", ")", "if", "self", ".", "_resource_manager", ":", "self", ".", "_resource_manager", ".", "_terminate_resource_request", "(", ")", "self", ".", "_prof", ".", "prof", "(", "'termination done'", ",", "uid", "=", "self", ".", "_uid", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_prof", ".", "prof", "(", "'start termination'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "exception", "(", "'Error in AppManager: %s'", "%", "ex", ")", "# Terminate threads in following order: wfp, helper, synchronizer", "if", "self", ".", "_wfp", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating WFprocessor'", ")", "self", ".", "_wfp", ".", "terminate_processor", "(", ")", "if", "self", ".", "_task_manager", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating task manager process'", ")", "self", ".", "_task_manager", ".", "terminate_manager", "(", ")", "self", ".", "_task_manager", ".", "terminate_heartbeat", "(", ")", "if", "self", ".", "_sync_thread", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating synchronizer thread'", ")", "self", ".", "_terminate_sync", ".", "set", "(", ")", "self", ".", "_sync_thread", ".", "join", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Synchronizer thread terminated'", ")", "if", "self", ".", "_resource_manager", ":", "self", ".", "_resource_manager", ".", "_terminate_resource_request", "(", ")", "self", ".", "_prof", ".", "prof", "(", "'termination done'", ",", "uid", "=", "self", ".", "_uid", ")", "raise" ]
**Purpose**: Run the application manager. Once the workflow and resource manager have been assigned. Invoking this method will start the setting up the communication infrastructure, submitting a resource request and then submission of all the tasks.
[ "**", "Purpose", "**", ":", "Run", "the", "application", "manager", ".", "Once", "the", "workflow", "and", "resource", "manager", "have", "been", "assigned", ".", "Invoking", "this", "method", "will", "start", "the", "setting", "up", "the", "communication", "infrastructure", "submitting", "a", "resource", "request", "and", "then", "submission", "of", "all", "the", "tasks", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/appmanager.py#L247-L503
radical-cybertools/radical.entk
src/radical/entk/appman/appmanager.py
AppManager._setup_mqs
def _setup_mqs(self): """ **Purpose**: Setup RabbitMQ system on the client side. We instantiate queue(s) 'pendingq-*' for communication between the enqueuer thread and the task manager process. We instantiate queue(s) 'completedq-*' for communication between the task manager and dequeuer thread. We instantiate queue 'sync-to-master' for communication from enqueuer/dequeuer/task_manager to the synchronizer thread. We instantiate queue 'sync-ack' for communication from synchronizer thread to enqueuer/dequeuer/task_manager. Details: All queues are durable: Even if the RabbitMQ server goes down, the queues are saved to disk and can be retrieved. This also means that after an erroneous run the queues might still have unacknowledged messages and will contain messages from that run. Hence, in every new run, we first delete the queue and create a new one. """ try: self._prof.prof('init mqs setup', uid=self._uid) self._logger.debug('Setting up mq connection and channel') mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() self._logger.debug('Connection and channel setup successful') self._logger.debug('Setting up all exchanges and queues') qs = [ '%s-tmgr-to-sync' % self._sid, '%s-cb-to-sync' % self._sid, '%s-enq-to-sync' % self._sid, '%s-deq-to-sync' % self._sid, '%s-sync-to-tmgr' % self._sid, '%s-sync-to-cb' % self._sid, '%s-sync-to-enq' % self._sid, '%s-sync-to-deq' % self._sid ] for i in range(1, self._num_pending_qs + 1): queue_name = '%s-pendingq-%s' % (self._sid, i) self._pending_queue.append(queue_name) qs.append(queue_name) for i in range(1, self._num_completed_qs + 1): queue_name = '%s-completedq-%s' % (self._sid, i) self._completed_queue.append(queue_name) qs.append(queue_name) f = open('.%s.txt' % self._sid, 'w') for q in qs: # Durable Qs will not be lost if rabbitmq server crashes mq_channel.queue_declare(queue=q) f.write(q + '\n') f.close() self._logger.debug('All exchanges and queues are setup') self._prof.prof('mqs setup done', uid=self._uid) return True except Exception, ex: self._logger.exception('Error setting RabbitMQ system: %s' % ex) raise
python
def _setup_mqs(self): """ **Purpose**: Setup RabbitMQ system on the client side. We instantiate queue(s) 'pendingq-*' for communication between the enqueuer thread and the task manager process. We instantiate queue(s) 'completedq-*' for communication between the task manager and dequeuer thread. We instantiate queue 'sync-to-master' for communication from enqueuer/dequeuer/task_manager to the synchronizer thread. We instantiate queue 'sync-ack' for communication from synchronizer thread to enqueuer/dequeuer/task_manager. Details: All queues are durable: Even if the RabbitMQ server goes down, the queues are saved to disk and can be retrieved. This also means that after an erroneous run the queues might still have unacknowledged messages and will contain messages from that run. Hence, in every new run, we first delete the queue and create a new one. """ try: self._prof.prof('init mqs setup', uid=self._uid) self._logger.debug('Setting up mq connection and channel') mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() self._logger.debug('Connection and channel setup successful') self._logger.debug('Setting up all exchanges and queues') qs = [ '%s-tmgr-to-sync' % self._sid, '%s-cb-to-sync' % self._sid, '%s-enq-to-sync' % self._sid, '%s-deq-to-sync' % self._sid, '%s-sync-to-tmgr' % self._sid, '%s-sync-to-cb' % self._sid, '%s-sync-to-enq' % self._sid, '%s-sync-to-deq' % self._sid ] for i in range(1, self._num_pending_qs + 1): queue_name = '%s-pendingq-%s' % (self._sid, i) self._pending_queue.append(queue_name) qs.append(queue_name) for i in range(1, self._num_completed_qs + 1): queue_name = '%s-completedq-%s' % (self._sid, i) self._completed_queue.append(queue_name) qs.append(queue_name) f = open('.%s.txt' % self._sid, 'w') for q in qs: # Durable Qs will not be lost if rabbitmq server crashes mq_channel.queue_declare(queue=q) f.write(q + '\n') f.close() self._logger.debug('All exchanges and queues are setup') self._prof.prof('mqs setup done', uid=self._uid) return True except Exception, ex: self._logger.exception('Error setting RabbitMQ system: %s' % ex) raise
[ "def", "_setup_mqs", "(", "self", ")", ":", "try", ":", "self", ".", "_prof", ".", "prof", "(", "'init mqs setup'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "debug", "(", "'Setting up mq connection and channel'", ")", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "self", ".", "_mq_hostname", ",", "port", "=", "self", ".", "_port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "self", ".", "_logger", ".", "debug", "(", "'Connection and channel setup successful'", ")", "self", ".", "_logger", ".", "debug", "(", "'Setting up all exchanges and queues'", ")", "qs", "=", "[", "'%s-tmgr-to-sync'", "%", "self", ".", "_sid", ",", "'%s-cb-to-sync'", "%", "self", ".", "_sid", ",", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ",", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ",", "'%s-sync-to-tmgr'", "%", "self", ".", "_sid", ",", "'%s-sync-to-cb'", "%", "self", ".", "_sid", ",", "'%s-sync-to-enq'", "%", "self", ".", "_sid", ",", "'%s-sync-to-deq'", "%", "self", ".", "_sid", "]", "for", "i", "in", "range", "(", "1", ",", "self", ".", "_num_pending_qs", "+", "1", ")", ":", "queue_name", "=", "'%s-pendingq-%s'", "%", "(", "self", ".", "_sid", ",", "i", ")", "self", ".", "_pending_queue", ".", "append", "(", "queue_name", ")", "qs", ".", "append", "(", "queue_name", ")", "for", "i", "in", "range", "(", "1", ",", "self", ".", "_num_completed_qs", "+", "1", ")", ":", "queue_name", "=", "'%s-completedq-%s'", "%", "(", "self", ".", "_sid", ",", "i", ")", "self", ".", "_completed_queue", ".", "append", "(", "queue_name", ")", "qs", ".", "append", "(", "queue_name", ")", "f", "=", "open", "(", "'.%s.txt'", "%", "self", ".", "_sid", ",", "'w'", ")", "for", "q", "in", "qs", ":", "# Durable Qs will not be lost if rabbitmq server crashes", "mq_channel", ".", "queue_declare", "(", "queue", "=", "q", ")", "f", ".", "write", "(", "q", "+", "'\\n'", ")", "f", ".", "close", "(", ")", "self", ".", "_logger", ".", "debug", "(", "'All exchanges and queues are setup'", ")", "self", ".", "_prof", ".", "prof", "(", "'mqs setup done'", ",", "uid", "=", "self", ".", "_uid", ")", "return", "True", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Error setting RabbitMQ system: %s'", "%", "ex", ")", "raise" ]
**Purpose**: Setup RabbitMQ system on the client side. We instantiate queue(s) 'pendingq-*' for communication between the enqueuer thread and the task manager process. We instantiate queue(s) 'completedq-*' for communication between the task manager and dequeuer thread. We instantiate queue 'sync-to-master' for communication from enqueuer/dequeuer/task_manager to the synchronizer thread. We instantiate queue 'sync-ack' for communication from synchronizer thread to enqueuer/dequeuer/task_manager. Details: All queues are durable: Even if the RabbitMQ server goes down, the queues are saved to disk and can be retrieved. This also means that after an erroneous run the queues might still have unacknowledged messages and will contain messages from that run. Hence, in every new run, we first delete the queue and create a new one.
[ "**", "Purpose", "**", ":", "Setup", "RabbitMQ", "system", "on", "the", "client", "side", ".", "We", "instantiate", "queue", "(", "s", ")", "pendingq", "-", "*", "for", "communication", "between", "the", "enqueuer", "thread", "and", "the", "task", "manager", "process", ".", "We", "instantiate", "queue", "(", "s", ")", "completedq", "-", "*", "for", "communication", "between", "the", "task", "manager", "and", "dequeuer", "thread", ".", "We", "instantiate", "queue", "sync", "-", "to", "-", "master", "for", "communication", "from", "enqueuer", "/", "dequeuer", "/", "task_manager", "to", "the", "synchronizer", "thread", ".", "We", "instantiate", "queue", "sync", "-", "ack", "for", "communication", "from", "synchronizer", "thread", "to", "enqueuer", "/", "dequeuer", "/", "task_manager", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/appmanager.py#L527-L589
radical-cybertools/radical.entk
src/radical/entk/appman/appmanager.py
AppManager._synchronizer
def _synchronizer(self): """ **Purpose**: Thread in the master process to keep the workflow data structure in appmanager up to date. We receive pipelines, stages and tasks objects directly. The respective object is updated in this master process. Details: Important to note that acknowledgements of the type channel.basic_ack() is an acknowledgement to the server that the msg was received. This is not to be confused with the Ack sent to the enqueuer/dequeuer/task_manager through the sync-ack queue. """ try: self._prof.prof('synchronizer started', uid=self._uid) self._logger.info('synchronizer thread started') def task_update(msg, reply_to, corr_id, mq_channel): completed_task = Task() completed_task.from_dict(msg['object']) self._logger.info('Received %s with state %s' % (completed_task.uid, completed_task.state)) found_task = False # Traverse the entire workflow to find the correct task for pipe in self._workflow: if not pipe.completed: if completed_task.parent_pipeline['uid'] == pipe.uid: for stage in pipe.stages: if completed_task.parent_stage['uid'] == stage.uid: for task in stage.tasks: if (completed_task.uid == task.uid)and(completed_task.state != task.state): task.state = str(completed_task.state) self._logger.debug('Found task %s with state %s' % (task.uid, task.state)) if completed_task.path: task.path = str(completed_task.path) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % task.uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) self._report.ok('Update: ') self._report.info('%s state: %s\n' % (task.luid, task.state)) found_task = True if not found_task: # If there was a Stage update, but the Stage was not found in any of the Pipelines. This # means that this was a Stage that was added during runtime and the AppManager does not # know about it. The current solution is going to be: add it to the workflow object in the # AppManager via the synchronizer. self._prof.prof('Adap: adding new task') self._logger.info('Adding new task %s to parent stage: %s' % (completed_task.uid, stage.uid)) stage.add_tasks(completed_task) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % completed_task.uid) self._prof.prof('Adap: added new task') self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) self._report.ok('Update: ') self._report.info('%s state: %s\n' % (completed_task.luid, completed_task.state)) def stage_update(msg, reply_to, corr_id, mq_channel): completed_stage = Stage() completed_stage.from_dict(msg['object']) self._logger.info('Received %s with state %s' % (completed_stage.uid, completed_stage.state)) found_stage = False # Traverse the entire workflow to find the correct stage for pipe in self._workflow: if not pipe.completed: if completed_stage.parent_pipeline['uid'] == pipe.uid: self._logger.info('Found parent pipeline: %s' % pipe.uid) for stage in pipe.stages: if (completed_stage.uid == stage.uid)and(completed_stage.state != stage.state): self._logger.debug('Found stage %s' % stage.uid) stage.state = str(completed_stage.state) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % stage.uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) self._report.ok('Update: ') self._report.info('%s state: %s\n' % (stage.luid, stage.state)) found_stage = True if not found_stage: # If there was a Stage update, but the Stage was not found in any of the Pipelines. This # means that this was a Stage that was added during runtime and the AppManager does not # know about it. The current solution is going to be: add it to the workflow object in the # AppManager via the synchronizer. self._prof.prof('Adap: adding new stage', uid=self._uid) self._logger.info('Adding new stage %s to parent pipeline: %s' % (completed_stage.uid, pipe.uid)) pipe.add_stages(completed_stage) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % completed_stage.uid) self._prof.prof('Adap: adding new stage', uid=self._uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) def pipeline_update(msg, reply_to, corr_id, mq_channel): completed_pipeline = Pipeline() completed_pipeline.from_dict(msg['object']) self._logger.info('Received %s with state %s' % (completed_pipeline.uid, completed_pipeline.state)) # Traverse the entire workflow to find the correct pipeline for pipe in self._workflow: if not pipe.completed: if (completed_pipeline.uid == pipe.uid)and(completed_pipeline.state != pipe.state): pipe.state = str(completed_pipeline.state) self._logger.info('Found pipeline %s, state %s, completed %s' % (pipe.uid, pipe.state, pipe.completed) ) # Reply with ack msg to the sender mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % pipe.uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) # Keep the assignment of the completed flag after sending the acknowledgment # back. Otherwise the MainThread takes lock over the pipeline because of logging # and profiling if completed_pipeline.completed: pipe._completed_flag.set() self._report.ok('Update: ') self._report.info('%s state: %s\n' % (pipe.luid, pipe.state)) mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() last = time.time() while not self._terminate_sync.is_set(): #------------------------------------------------------------------------------------------------------- # Messages between tmgr Main thread and synchronizer -- only Task objects method_frame, props, body = mq_channel.basic_get(queue='%s-tmgr-to-sync' % self._sid) """ The message received is a JSON object with the following structure: msg = { 'type': 'Pipeline'/'Stage'/'Task', 'object': json/dict } """ if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-tmgr' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- #------------------------------------------------------------------------------------------------------- # Messages between callback thread and synchronizer -- only Task objects method_frame, props, body = mq_channel.basic_get(queue='%s-cb-to-sync' % self._sid) """ The message received is a JSON object with the following structure: msg = { 'type': 'Pipeline'/'Stage'/'Task', 'object': json/dict } """ if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-cb' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- #------------------------------------------------------------------------------------------------------- # Messages between enqueue thread and synchronizer -- Task, Stage or Pipeline method_frame, props, body = mq_channel.basic_get(queue='%s-enq-to-sync' % self._sid) if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-enq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Stage': stage_update(msg, '%s-sync-to-enq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Pipeline': pipeline_update(msg, '%s-sync-to-enq' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- #------------------------------------------------------------------------------------------------------- # Messages between dequeue thread and synchronizer -- Task, Stage or Pipeline method_frame, props, body = mq_channel.basic_get(queue='%s-deq-to-sync' % self._sid) if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-deq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Stage': stage_update(msg, '%s-sync-to-deq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Pipeline': pipeline_update(msg, '%s-sync-to-deq' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- # Appease pika cos it thinks the connection is dead now = time.time() if now - last >= self._rmq_ping_interval: mq_connection.process_data_events() last = now self._prof.prof('terminating synchronizer', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to terminate synchronizer thread gracefully...') raise KeyboardInterrupt except Exception, ex: self._logger.exception('Unknown error in synchronizer: %s. \n Terminating thread' % ex) raise
python
def _synchronizer(self): """ **Purpose**: Thread in the master process to keep the workflow data structure in appmanager up to date. We receive pipelines, stages and tasks objects directly. The respective object is updated in this master process. Details: Important to note that acknowledgements of the type channel.basic_ack() is an acknowledgement to the server that the msg was received. This is not to be confused with the Ack sent to the enqueuer/dequeuer/task_manager through the sync-ack queue. """ try: self._prof.prof('synchronizer started', uid=self._uid) self._logger.info('synchronizer thread started') def task_update(msg, reply_to, corr_id, mq_channel): completed_task = Task() completed_task.from_dict(msg['object']) self._logger.info('Received %s with state %s' % (completed_task.uid, completed_task.state)) found_task = False # Traverse the entire workflow to find the correct task for pipe in self._workflow: if not pipe.completed: if completed_task.parent_pipeline['uid'] == pipe.uid: for stage in pipe.stages: if completed_task.parent_stage['uid'] == stage.uid: for task in stage.tasks: if (completed_task.uid == task.uid)and(completed_task.state != task.state): task.state = str(completed_task.state) self._logger.debug('Found task %s with state %s' % (task.uid, task.state)) if completed_task.path: task.path = str(completed_task.path) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % task.uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) self._report.ok('Update: ') self._report.info('%s state: %s\n' % (task.luid, task.state)) found_task = True if not found_task: # If there was a Stage update, but the Stage was not found in any of the Pipelines. This # means that this was a Stage that was added during runtime and the AppManager does not # know about it. The current solution is going to be: add it to the workflow object in the # AppManager via the synchronizer. self._prof.prof('Adap: adding new task') self._logger.info('Adding new task %s to parent stage: %s' % (completed_task.uid, stage.uid)) stage.add_tasks(completed_task) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % completed_task.uid) self._prof.prof('Adap: added new task') self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) self._report.ok('Update: ') self._report.info('%s state: %s\n' % (completed_task.luid, completed_task.state)) def stage_update(msg, reply_to, corr_id, mq_channel): completed_stage = Stage() completed_stage.from_dict(msg['object']) self._logger.info('Received %s with state %s' % (completed_stage.uid, completed_stage.state)) found_stage = False # Traverse the entire workflow to find the correct stage for pipe in self._workflow: if not pipe.completed: if completed_stage.parent_pipeline['uid'] == pipe.uid: self._logger.info('Found parent pipeline: %s' % pipe.uid) for stage in pipe.stages: if (completed_stage.uid == stage.uid)and(completed_stage.state != stage.state): self._logger.debug('Found stage %s' % stage.uid) stage.state = str(completed_stage.state) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % stage.uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) self._report.ok('Update: ') self._report.info('%s state: %s\n' % (stage.luid, stage.state)) found_stage = True if not found_stage: # If there was a Stage update, but the Stage was not found in any of the Pipelines. This # means that this was a Stage that was added during runtime and the AppManager does not # know about it. The current solution is going to be: add it to the workflow object in the # AppManager via the synchronizer. self._prof.prof('Adap: adding new stage', uid=self._uid) self._logger.info('Adding new stage %s to parent pipeline: %s' % (completed_stage.uid, pipe.uid)) pipe.add_stages(completed_stage) mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % completed_stage.uid) self._prof.prof('Adap: adding new stage', uid=self._uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) def pipeline_update(msg, reply_to, corr_id, mq_channel): completed_pipeline = Pipeline() completed_pipeline.from_dict(msg['object']) self._logger.info('Received %s with state %s' % (completed_pipeline.uid, completed_pipeline.state)) # Traverse the entire workflow to find the correct pipeline for pipe in self._workflow: if not pipe.completed: if (completed_pipeline.uid == pipe.uid)and(completed_pipeline.state != pipe.state): pipe.state = str(completed_pipeline.state) self._logger.info('Found pipeline %s, state %s, completed %s' % (pipe.uid, pipe.state, pipe.completed) ) # Reply with ack msg to the sender mq_channel.basic_publish(exchange='', routing_key=reply_to, properties=pika.BasicProperties( correlation_id=corr_id), body='%s-ack' % pipe.uid) self._prof.prof('publishing sync ack for obj with state %s' % msg['object']['state'], uid=msg['object']['uid'] ) mq_channel.basic_ack(delivery_tag=method_frame.delivery_tag) # Keep the assignment of the completed flag after sending the acknowledgment # back. Otherwise the MainThread takes lock over the pipeline because of logging # and profiling if completed_pipeline.completed: pipe._completed_flag.set() self._report.ok('Update: ') self._report.info('%s state: %s\n' % (pipe.luid, pipe.state)) mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() last = time.time() while not self._terminate_sync.is_set(): #------------------------------------------------------------------------------------------------------- # Messages between tmgr Main thread and synchronizer -- only Task objects method_frame, props, body = mq_channel.basic_get(queue='%s-tmgr-to-sync' % self._sid) """ The message received is a JSON object with the following structure: msg = { 'type': 'Pipeline'/'Stage'/'Task', 'object': json/dict } """ if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-tmgr' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- #------------------------------------------------------------------------------------------------------- # Messages between callback thread and synchronizer -- only Task objects method_frame, props, body = mq_channel.basic_get(queue='%s-cb-to-sync' % self._sid) """ The message received is a JSON object with the following structure: msg = { 'type': 'Pipeline'/'Stage'/'Task', 'object': json/dict } """ if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-cb' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- #------------------------------------------------------------------------------------------------------- # Messages between enqueue thread and synchronizer -- Task, Stage or Pipeline method_frame, props, body = mq_channel.basic_get(queue='%s-enq-to-sync' % self._sid) if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-enq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Stage': stage_update(msg, '%s-sync-to-enq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Pipeline': pipeline_update(msg, '%s-sync-to-enq' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- #------------------------------------------------------------------------------------------------------- # Messages between dequeue thread and synchronizer -- Task, Stage or Pipeline method_frame, props, body = mq_channel.basic_get(queue='%s-deq-to-sync' % self._sid) if body: msg = json.loads(body) self._prof.prof('received obj with state %s for sync' % msg['object']['state'], uid=msg['object']['uid']) self._logger.debug('received %s with state %s for sync' % (msg['object']['uid'], msg['object']['state'])) if msg['type'] == 'Task': task_update(msg, '%s-sync-to-deq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Stage': stage_update(msg, '%s-sync-to-deq' % self._sid, props.correlation_id, mq_channel) elif msg['type'] == 'Pipeline': pipeline_update(msg, '%s-sync-to-deq' % self._sid, props.correlation_id, mq_channel) #------------------------------------------------------------------------------------------------------- # Appease pika cos it thinks the connection is dead now = time.time() if now - last >= self._rmq_ping_interval: mq_connection.process_data_events() last = now self._prof.prof('terminating synchronizer', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to terminate synchronizer thread gracefully...') raise KeyboardInterrupt except Exception, ex: self._logger.exception('Unknown error in synchronizer: %s. \n Terminating thread' % ex) raise
[ "def", "_synchronizer", "(", "self", ")", ":", "try", ":", "self", ".", "_prof", ".", "prof", "(", "'synchronizer started'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'synchronizer thread started'", ")", "def", "task_update", "(", "msg", ",", "reply_to", ",", "corr_id", ",", "mq_channel", ")", ":", "completed_task", "=", "Task", "(", ")", "completed_task", ".", "from_dict", "(", "msg", "[", "'object'", "]", ")", "self", ".", "_logger", ".", "info", "(", "'Received %s with state %s'", "%", "(", "completed_task", ".", "uid", ",", "completed_task", ".", "state", ")", ")", "found_task", "=", "False", "# Traverse the entire workflow to find the correct task", "for", "pipe", "in", "self", ".", "_workflow", ":", "if", "not", "pipe", ".", "completed", ":", "if", "completed_task", ".", "parent_pipeline", "[", "'uid'", "]", "==", "pipe", ".", "uid", ":", "for", "stage", "in", "pipe", ".", "stages", ":", "if", "completed_task", ".", "parent_stage", "[", "'uid'", "]", "==", "stage", ".", "uid", ":", "for", "task", "in", "stage", ".", "tasks", ":", "if", "(", "completed_task", ".", "uid", "==", "task", ".", "uid", ")", "and", "(", "completed_task", ".", "state", "!=", "task", ".", "state", ")", ":", "task", ".", "state", "=", "str", "(", "completed_task", ".", "state", ")", "self", ".", "_logger", ".", "debug", "(", "'Found task %s with state %s'", "%", "(", "task", ".", "uid", ",", "task", ".", "state", ")", ")", "if", "completed_task", ".", "path", ":", "task", ".", "path", "=", "str", "(", "completed_task", ".", "path", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "reply_to", ",", "properties", "=", "pika", ".", "BasicProperties", "(", "correlation_id", "=", "corr_id", ")", ",", "body", "=", "'%s-ack'", "%", "task", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'publishing sync ack for obj with state %s'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "self", ".", "_report", ".", "ok", "(", "'Update: '", ")", "self", ".", "_report", ".", "info", "(", "'%s state: %s\\n'", "%", "(", "task", ".", "luid", ",", "task", ".", "state", ")", ")", "found_task", "=", "True", "if", "not", "found_task", ":", "# If there was a Stage update, but the Stage was not found in any of the Pipelines. This", "# means that this was a Stage that was added during runtime and the AppManager does not", "# know about it. The current solution is going to be: add it to the workflow object in the", "# AppManager via the synchronizer.", "self", ".", "_prof", ".", "prof", "(", "'Adap: adding new task'", ")", "self", ".", "_logger", ".", "info", "(", "'Adding new task %s to parent stage: %s'", "%", "(", "completed_task", ".", "uid", ",", "stage", ".", "uid", ")", ")", "stage", ".", "add_tasks", "(", "completed_task", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "reply_to", ",", "properties", "=", "pika", ".", "BasicProperties", "(", "correlation_id", "=", "corr_id", ")", ",", "body", "=", "'%s-ack'", "%", "completed_task", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'Adap: added new task'", ")", "self", ".", "_prof", ".", "prof", "(", "'publishing sync ack for obj with state %s'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "self", ".", "_report", ".", "ok", "(", "'Update: '", ")", "self", ".", "_report", ".", "info", "(", "'%s state: %s\\n'", "%", "(", "completed_task", ".", "luid", ",", "completed_task", ".", "state", ")", ")", "def", "stage_update", "(", "msg", ",", "reply_to", ",", "corr_id", ",", "mq_channel", ")", ":", "completed_stage", "=", "Stage", "(", ")", "completed_stage", ".", "from_dict", "(", "msg", "[", "'object'", "]", ")", "self", ".", "_logger", ".", "info", "(", "'Received %s with state %s'", "%", "(", "completed_stage", ".", "uid", ",", "completed_stage", ".", "state", ")", ")", "found_stage", "=", "False", "# Traverse the entire workflow to find the correct stage", "for", "pipe", "in", "self", ".", "_workflow", ":", "if", "not", "pipe", ".", "completed", ":", "if", "completed_stage", ".", "parent_pipeline", "[", "'uid'", "]", "==", "pipe", ".", "uid", ":", "self", ".", "_logger", ".", "info", "(", "'Found parent pipeline: %s'", "%", "pipe", ".", "uid", ")", "for", "stage", "in", "pipe", ".", "stages", ":", "if", "(", "completed_stage", ".", "uid", "==", "stage", ".", "uid", ")", "and", "(", "completed_stage", ".", "state", "!=", "stage", ".", "state", ")", ":", "self", ".", "_logger", ".", "debug", "(", "'Found stage %s'", "%", "stage", ".", "uid", ")", "stage", ".", "state", "=", "str", "(", "completed_stage", ".", "state", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "reply_to", ",", "properties", "=", "pika", ".", "BasicProperties", "(", "correlation_id", "=", "corr_id", ")", ",", "body", "=", "'%s-ack'", "%", "stage", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'publishing sync ack for obj with state %s'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "self", ".", "_report", ".", "ok", "(", "'Update: '", ")", "self", ".", "_report", ".", "info", "(", "'%s state: %s\\n'", "%", "(", "stage", ".", "luid", ",", "stage", ".", "state", ")", ")", "found_stage", "=", "True", "if", "not", "found_stage", ":", "# If there was a Stage update, but the Stage was not found in any of the Pipelines. This", "# means that this was a Stage that was added during runtime and the AppManager does not", "# know about it. The current solution is going to be: add it to the workflow object in the", "# AppManager via the synchronizer.", "self", ".", "_prof", ".", "prof", "(", "'Adap: adding new stage'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'Adding new stage %s to parent pipeline: %s'", "%", "(", "completed_stage", ".", "uid", ",", "pipe", ".", "uid", ")", ")", "pipe", ".", "add_stages", "(", "completed_stage", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "reply_to", ",", "properties", "=", "pika", ".", "BasicProperties", "(", "correlation_id", "=", "corr_id", ")", ",", "body", "=", "'%s-ack'", "%", "completed_stage", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'Adap: adding new stage'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_prof", ".", "prof", "(", "'publishing sync ack for obj with state %s'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "def", "pipeline_update", "(", "msg", ",", "reply_to", ",", "corr_id", ",", "mq_channel", ")", ":", "completed_pipeline", "=", "Pipeline", "(", ")", "completed_pipeline", ".", "from_dict", "(", "msg", "[", "'object'", "]", ")", "self", ".", "_logger", ".", "info", "(", "'Received %s with state %s'", "%", "(", "completed_pipeline", ".", "uid", ",", "completed_pipeline", ".", "state", ")", ")", "# Traverse the entire workflow to find the correct pipeline", "for", "pipe", "in", "self", ".", "_workflow", ":", "if", "not", "pipe", ".", "completed", ":", "if", "(", "completed_pipeline", ".", "uid", "==", "pipe", ".", "uid", ")", "and", "(", "completed_pipeline", ".", "state", "!=", "pipe", ".", "state", ")", ":", "pipe", ".", "state", "=", "str", "(", "completed_pipeline", ".", "state", ")", "self", ".", "_logger", ".", "info", "(", "'Found pipeline %s, state %s, completed %s'", "%", "(", "pipe", ".", "uid", ",", "pipe", ".", "state", ",", "pipe", ".", "completed", ")", ")", "# Reply with ack msg to the sender", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "reply_to", ",", "properties", "=", "pika", ".", "BasicProperties", "(", "correlation_id", "=", "corr_id", ")", ",", "body", "=", "'%s-ack'", "%", "pipe", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'publishing sync ack for obj with state %s'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "# Keep the assignment of the completed flag after sending the acknowledgment", "# back. Otherwise the MainThread takes lock over the pipeline because of logging", "# and profiling", "if", "completed_pipeline", ".", "completed", ":", "pipe", ".", "_completed_flag", ".", "set", "(", ")", "self", ".", "_report", ".", "ok", "(", "'Update: '", ")", "self", ".", "_report", ".", "info", "(", "'%s state: %s\\n'", "%", "(", "pipe", ".", "luid", ",", "pipe", ".", "state", ")", ")", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "self", ".", "_mq_hostname", ",", "port", "=", "self", ".", "_port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "last", "=", "time", ".", "time", "(", ")", "while", "not", "self", ".", "_terminate_sync", ".", "is_set", "(", ")", ":", "#-------------------------------------------------------------------------------------------------------", "# Messages between tmgr Main thread and synchronizer -- only Task objects", "method_frame", ",", "props", ",", "body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "'%s-tmgr-to-sync'", "%", "self", ".", "_sid", ")", "\"\"\"\n The message received is a JSON object with the following structure:\n\n msg = {\n 'type': 'Pipeline'/'Stage'/'Task',\n 'object': json/dict\n }\n \"\"\"", "if", "body", ":", "msg", "=", "json", ".", "loads", "(", "body", ")", "self", ".", "_prof", ".", "prof", "(", "'received obj with state %s for sync'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "self", ".", "_logger", ".", "debug", "(", "'received %s with state %s for sync'", "%", "(", "msg", "[", "'object'", "]", "[", "'uid'", "]", ",", "msg", "[", "'object'", "]", "[", "'state'", "]", ")", ")", "if", "msg", "[", "'type'", "]", "==", "'Task'", ":", "task_update", "(", "msg", ",", "'%s-sync-to-tmgr'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "#-------------------------------------------------------------------------------------------------------", "#-------------------------------------------------------------------------------------------------------", "# Messages between callback thread and synchronizer -- only Task objects", "method_frame", ",", "props", ",", "body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "'%s-cb-to-sync'", "%", "self", ".", "_sid", ")", "\"\"\"\n The message received is a JSON object with the following structure:\n\n msg = {\n 'type': 'Pipeline'/'Stage'/'Task',\n 'object': json/dict\n }\n \"\"\"", "if", "body", ":", "msg", "=", "json", ".", "loads", "(", "body", ")", "self", ".", "_prof", ".", "prof", "(", "'received obj with state %s for sync'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "self", ".", "_logger", ".", "debug", "(", "'received %s with state %s for sync'", "%", "(", "msg", "[", "'object'", "]", "[", "'uid'", "]", ",", "msg", "[", "'object'", "]", "[", "'state'", "]", ")", ")", "if", "msg", "[", "'type'", "]", "==", "'Task'", ":", "task_update", "(", "msg", ",", "'%s-sync-to-cb'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "#-------------------------------------------------------------------------------------------------------", "#-------------------------------------------------------------------------------------------------------", "# Messages between enqueue thread and synchronizer -- Task, Stage or Pipeline", "method_frame", ",", "props", ",", "body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ")", "if", "body", ":", "msg", "=", "json", ".", "loads", "(", "body", ")", "self", ".", "_prof", ".", "prof", "(", "'received obj with state %s for sync'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "self", ".", "_logger", ".", "debug", "(", "'received %s with state %s for sync'", "%", "(", "msg", "[", "'object'", "]", "[", "'uid'", "]", ",", "msg", "[", "'object'", "]", "[", "'state'", "]", ")", ")", "if", "msg", "[", "'type'", "]", "==", "'Task'", ":", "task_update", "(", "msg", ",", "'%s-sync-to-enq'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "elif", "msg", "[", "'type'", "]", "==", "'Stage'", ":", "stage_update", "(", "msg", ",", "'%s-sync-to-enq'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "elif", "msg", "[", "'type'", "]", "==", "'Pipeline'", ":", "pipeline_update", "(", "msg", ",", "'%s-sync-to-enq'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "#-------------------------------------------------------------------------------------------------------", "#-------------------------------------------------------------------------------------------------------", "# Messages between dequeue thread and synchronizer -- Task, Stage or Pipeline", "method_frame", ",", "props", ",", "body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ")", "if", "body", ":", "msg", "=", "json", ".", "loads", "(", "body", ")", "self", ".", "_prof", ".", "prof", "(", "'received obj with state %s for sync'", "%", "msg", "[", "'object'", "]", "[", "'state'", "]", ",", "uid", "=", "msg", "[", "'object'", "]", "[", "'uid'", "]", ")", "self", ".", "_logger", ".", "debug", "(", "'received %s with state %s for sync'", "%", "(", "msg", "[", "'object'", "]", "[", "'uid'", "]", ",", "msg", "[", "'object'", "]", "[", "'state'", "]", ")", ")", "if", "msg", "[", "'type'", "]", "==", "'Task'", ":", "task_update", "(", "msg", ",", "'%s-sync-to-deq'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "elif", "msg", "[", "'type'", "]", "==", "'Stage'", ":", "stage_update", "(", "msg", ",", "'%s-sync-to-deq'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "elif", "msg", "[", "'type'", "]", "==", "'Pipeline'", ":", "pipeline_update", "(", "msg", ",", "'%s-sync-to-deq'", "%", "self", ".", "_sid", ",", "props", ".", "correlation_id", ",", "mq_channel", ")", "#-------------------------------------------------------------------------------------------------------", "# Appease pika cos it thinks the connection is dead", "now", "=", "time", ".", "time", "(", ")", "if", "now", "-", "last", ">=", "self", ".", "_rmq_ping_interval", ":", "mq_connection", ".", "process_data_events", "(", ")", "last", "=", "now", "self", ".", "_prof", ".", "prof", "(", "'terminating synchronizer'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "KeyboardInterrupt", ":", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to terminate synchronizer thread gracefully...'", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Unknown error in synchronizer: %s. \\n Terminating thread'", "%", "ex", ")", "raise" ]
**Purpose**: Thread in the master process to keep the workflow data structure in appmanager up to date. We receive pipelines, stages and tasks objects directly. The respective object is updated in this master process. Details: Important to note that acknowledgements of the type channel.basic_ack() is an acknowledgement to the server that the msg was received. This is not to be confused with the Ack sent to the enqueuer/dequeuer/task_manager through the sync-ack queue.
[ "**", "Purpose", "**", ":", "Thread", "in", "the", "master", "process", "to", "keep", "the", "workflow", "data", "structure", "in", "appmanager", "up", "to", "date", ".", "We", "receive", "pipelines", "stages", "and", "tasks", "objects", "directly", ".", "The", "respective", "object", "is", "updated", "in", "this", "master", "process", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/appmanager.py#L619-L958
bradmontgomery/django-redis-metrics
redis_metrics/forms.py
MetricCategoryForm.categorize_metrics
def categorize_metrics(self): """Called only on a valid form, this method will place the chosen metrics in the given catgory.""" category = self.cleaned_data['category_name'] metrics = self.cleaned_data['metrics'] self.r.reset_category(category, metrics)
python
def categorize_metrics(self): """Called only on a valid form, this method will place the chosen metrics in the given catgory.""" category = self.cleaned_data['category_name'] metrics = self.cleaned_data['metrics'] self.r.reset_category(category, metrics)
[ "def", "categorize_metrics", "(", "self", ")", ":", "category", "=", "self", ".", "cleaned_data", "[", "'category_name'", "]", "metrics", "=", "self", ".", "cleaned_data", "[", "'metrics'", "]", "self", ".", "r", ".", "reset_category", "(", "category", ",", "metrics", ")" ]
Called only on a valid form, this method will place the chosen metrics in the given catgory.
[ "Called", "only", "on", "a", "valid", "form", "this", "method", "will", "place", "the", "chosen", "metrics", "in", "the", "given", "catgory", "." ]
train
https://github.com/bradmontgomery/django-redis-metrics/blob/2c92332920113d28c39234b949aa496b39a091d1/redis_metrics/forms.py#L50-L55
radical-cybertools/radical.entk
src/radical/entk/execman/rp/resource_manager.py
ResourceManager._submit_resource_request
def _submit_resource_request(self): """ **Purpose**: Create and submits a RADICAL Pilot Job as per the user provided resource description """ try: self._prof.prof('creating rreq', uid=self._uid) def _pilot_state_cb(pilot, state): self._logger.info('Pilot %s state: %s' % (pilot.uid, state)) if state == rp.FAILED: self._logger.error('Pilot has failed') elif state == rp.DONE: self._logger.error('Pilot has completed') self._session = rp.Session(dburl=self._mlab_url, uid=self._sid) self._pmgr = rp.PilotManager(session=self._session) self._pmgr.register_callback(_pilot_state_cb) pd_init = { 'resource': self._resource, 'runtime': self._walltime, 'cores': self._cpus, 'project': self._project, } if self._gpus: pd_init['gpus'] = self._gpus if self._access_schema: pd_init['access_schema'] = self._access_schema if self._queue: pd_init['queue'] = self._queue if self._rts_config.get('sandbox_cleanup', None): pd_init['cleanup'] = True # Create Compute Pilot with validated resource description pdesc = rp.ComputePilotDescription(pd_init) self._prof.prof('rreq created', uid=self._uid) # Launch the pilot self._pilot = self._pmgr.submit_pilots(pdesc) self._prof.prof('rreq submitted', uid=self._uid) shared_staging_directives = list() for data in self._shared_data: temp = { 'source': data, 'target': 'pilot:///' + os.path.basename(data) } shared_staging_directives.append(temp) self._pilot.stage_in(shared_staging_directives) self._prof.prof('shared data staging initiated', uid=self._uid) self._logger.info('Resource request submission successful.. waiting for pilot to go Active') # Wait for pilot to go active self._pilot.wait([rp.PMGR_ACTIVE, rp.FAILED, rp.CANCELED]) self._prof.prof('resource active', uid=self._uid) self._logger.info('Pilot is now active') except KeyboardInterrupt: if self._session: self._session.close() self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit callback thread gracefully...') raise KeyboardInterrupt except Exception, ex: self._logger.exception('Resource request submission failed') raise
python
def _submit_resource_request(self): """ **Purpose**: Create and submits a RADICAL Pilot Job as per the user provided resource description """ try: self._prof.prof('creating rreq', uid=self._uid) def _pilot_state_cb(pilot, state): self._logger.info('Pilot %s state: %s' % (pilot.uid, state)) if state == rp.FAILED: self._logger.error('Pilot has failed') elif state == rp.DONE: self._logger.error('Pilot has completed') self._session = rp.Session(dburl=self._mlab_url, uid=self._sid) self._pmgr = rp.PilotManager(session=self._session) self._pmgr.register_callback(_pilot_state_cb) pd_init = { 'resource': self._resource, 'runtime': self._walltime, 'cores': self._cpus, 'project': self._project, } if self._gpus: pd_init['gpus'] = self._gpus if self._access_schema: pd_init['access_schema'] = self._access_schema if self._queue: pd_init['queue'] = self._queue if self._rts_config.get('sandbox_cleanup', None): pd_init['cleanup'] = True # Create Compute Pilot with validated resource description pdesc = rp.ComputePilotDescription(pd_init) self._prof.prof('rreq created', uid=self._uid) # Launch the pilot self._pilot = self._pmgr.submit_pilots(pdesc) self._prof.prof('rreq submitted', uid=self._uid) shared_staging_directives = list() for data in self._shared_data: temp = { 'source': data, 'target': 'pilot:///' + os.path.basename(data) } shared_staging_directives.append(temp) self._pilot.stage_in(shared_staging_directives) self._prof.prof('shared data staging initiated', uid=self._uid) self._logger.info('Resource request submission successful.. waiting for pilot to go Active') # Wait for pilot to go active self._pilot.wait([rp.PMGR_ACTIVE, rp.FAILED, rp.CANCELED]) self._prof.prof('resource active', uid=self._uid) self._logger.info('Pilot is now active') except KeyboardInterrupt: if self._session: self._session.close() self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit callback thread gracefully...') raise KeyboardInterrupt except Exception, ex: self._logger.exception('Resource request submission failed') raise
[ "def", "_submit_resource_request", "(", "self", ")", ":", "try", ":", "self", ".", "_prof", ".", "prof", "(", "'creating rreq'", ",", "uid", "=", "self", ".", "_uid", ")", "def", "_pilot_state_cb", "(", "pilot", ",", "state", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Pilot %s state: %s'", "%", "(", "pilot", ".", "uid", ",", "state", ")", ")", "if", "state", "==", "rp", ".", "FAILED", ":", "self", ".", "_logger", ".", "error", "(", "'Pilot has failed'", ")", "elif", "state", "==", "rp", ".", "DONE", ":", "self", ".", "_logger", ".", "error", "(", "'Pilot has completed'", ")", "self", ".", "_session", "=", "rp", ".", "Session", "(", "dburl", "=", "self", ".", "_mlab_url", ",", "uid", "=", "self", ".", "_sid", ")", "self", ".", "_pmgr", "=", "rp", ".", "PilotManager", "(", "session", "=", "self", ".", "_session", ")", "self", ".", "_pmgr", ".", "register_callback", "(", "_pilot_state_cb", ")", "pd_init", "=", "{", "'resource'", ":", "self", ".", "_resource", ",", "'runtime'", ":", "self", ".", "_walltime", ",", "'cores'", ":", "self", ".", "_cpus", ",", "'project'", ":", "self", ".", "_project", ",", "}", "if", "self", ".", "_gpus", ":", "pd_init", "[", "'gpus'", "]", "=", "self", ".", "_gpus", "if", "self", ".", "_access_schema", ":", "pd_init", "[", "'access_schema'", "]", "=", "self", ".", "_access_schema", "if", "self", ".", "_queue", ":", "pd_init", "[", "'queue'", "]", "=", "self", ".", "_queue", "if", "self", ".", "_rts_config", ".", "get", "(", "'sandbox_cleanup'", ",", "None", ")", ":", "pd_init", "[", "'cleanup'", "]", "=", "True", "# Create Compute Pilot with validated resource description", "pdesc", "=", "rp", ".", "ComputePilotDescription", "(", "pd_init", ")", "self", ".", "_prof", ".", "prof", "(", "'rreq created'", ",", "uid", "=", "self", ".", "_uid", ")", "# Launch the pilot", "self", ".", "_pilot", "=", "self", ".", "_pmgr", ".", "submit_pilots", "(", "pdesc", ")", "self", ".", "_prof", ".", "prof", "(", "'rreq submitted'", ",", "uid", "=", "self", ".", "_uid", ")", "shared_staging_directives", "=", "list", "(", ")", "for", "data", "in", "self", ".", "_shared_data", ":", "temp", "=", "{", "'source'", ":", "data", ",", "'target'", ":", "'pilot:///'", "+", "os", ".", "path", ".", "basename", "(", "data", ")", "}", "shared_staging_directives", ".", "append", "(", "temp", ")", "self", ".", "_pilot", ".", "stage_in", "(", "shared_staging_directives", ")", "self", ".", "_prof", ".", "prof", "(", "'shared data staging initiated'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'Resource request submission successful.. waiting for pilot to go Active'", ")", "# Wait for pilot to go active", "self", ".", "_pilot", ".", "wait", "(", "[", "rp", ".", "PMGR_ACTIVE", ",", "rp", ".", "FAILED", ",", "rp", ".", "CANCELED", "]", ")", "self", ".", "_prof", ".", "prof", "(", "'resource active'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'Pilot is now active'", ")", "except", "KeyboardInterrupt", ":", "if", "self", ".", "_session", ":", "self", ".", "_session", ".", "close", "(", ")", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to exit callback thread gracefully...'", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Resource request submission failed'", ")", "raise" ]
**Purpose**: Create and submits a RADICAL Pilot Job as per the user provided resource description
[ "**", "Purpose", "**", ":", "Create", "and", "submits", "a", "RADICAL", "Pilot", "Job", "as", "per", "the", "user", "provided", "resource", "description" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/resource_manager.py#L106-L188
radical-cybertools/radical.entk
src/radical/entk/execman/rp/resource_manager.py
ResourceManager._terminate_resource_request
def _terminate_resource_request(self): """ **Purpose**: Cancel the RADICAL Pilot Job """ try: if self._pilot: self._prof.prof('canceling resource allocation', uid=self._uid) self._pilot.cancel() download_rp_profile = os.environ.get('RADICAL_PILOT_PROFILE', False) self._session.close(cleanup=self._rts_config.get('db_cleanup', False), download=download_rp_profile) self._prof.prof('resource allocation canceled', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit callback thread gracefully...') raise KeyboardInterrupt except Exception, ex: self._logger.exception('Could not cancel resource request, error: %s' % ex) raise
python
def _terminate_resource_request(self): """ **Purpose**: Cancel the RADICAL Pilot Job """ try: if self._pilot: self._prof.prof('canceling resource allocation', uid=self._uid) self._pilot.cancel() download_rp_profile = os.environ.get('RADICAL_PILOT_PROFILE', False) self._session.close(cleanup=self._rts_config.get('db_cleanup', False), download=download_rp_profile) self._prof.prof('resource allocation canceled', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit callback thread gracefully...') raise KeyboardInterrupt except Exception, ex: self._logger.exception('Could not cancel resource request, error: %s' % ex) raise
[ "def", "_terminate_resource_request", "(", "self", ")", ":", "try", ":", "if", "self", ".", "_pilot", ":", "self", ".", "_prof", ".", "prof", "(", "'canceling resource allocation'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_pilot", ".", "cancel", "(", ")", "download_rp_profile", "=", "os", ".", "environ", ".", "get", "(", "'RADICAL_PILOT_PROFILE'", ",", "False", ")", "self", ".", "_session", ".", "close", "(", "cleanup", "=", "self", ".", "_rts_config", ".", "get", "(", "'db_cleanup'", ",", "False", ")", ",", "download", "=", "download_rp_profile", ")", "self", ".", "_prof", ".", "prof", "(", "'resource allocation canceled'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "KeyboardInterrupt", ":", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to exit callback thread gracefully...'", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Could not cancel resource request, error: %s'", "%", "ex", ")", "raise" ]
**Purpose**: Cancel the RADICAL Pilot Job
[ "**", "Purpose", "**", ":", "Cancel", "the", "RADICAL", "Pilot", "Job" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/resource_manager.py#L190-L214
bertouttier/solaredge
solaredge/solaredge.py
Solaredge.get_list
def get_list(self, size=100, startIndex=0, searchText="", sortProperty="", sortOrder='ASC', status='Active,Pending'): """ Request service locations Returns ------- dict """ url = urljoin(BASEURL, "sites", "list") params = { 'api_key': self.token, 'size': size, 'startIndex': startIndex, 'sortOrder': sortOrder, 'status': status } if searchText: params['searchText'] = searchText if sortProperty: params['sortProperty'] = sortProperty r = requests.get(url, params) r.raise_for_status() return r.json()
python
def get_list(self, size=100, startIndex=0, searchText="", sortProperty="", sortOrder='ASC', status='Active,Pending'): """ Request service locations Returns ------- dict """ url = urljoin(BASEURL, "sites", "list") params = { 'api_key': self.token, 'size': size, 'startIndex': startIndex, 'sortOrder': sortOrder, 'status': status } if searchText: params['searchText'] = searchText if sortProperty: params['sortProperty'] = sortProperty r = requests.get(url, params) r.raise_for_status() return r.json()
[ "def", "get_list", "(", "self", ",", "size", "=", "100", ",", "startIndex", "=", "0", ",", "searchText", "=", "\"\"", ",", "sortProperty", "=", "\"\"", ",", "sortOrder", "=", "'ASC'", ",", "status", "=", "'Active,Pending'", ")", ":", "url", "=", "urljoin", "(", "BASEURL", ",", "\"sites\"", ",", "\"list\"", ")", "params", "=", "{", "'api_key'", ":", "self", ".", "token", ",", "'size'", ":", "size", ",", "'startIndex'", ":", "startIndex", ",", "'sortOrder'", ":", "sortOrder", ",", "'status'", ":", "status", "}", "if", "searchText", ":", "params", "[", "'searchText'", "]", "=", "searchText", "if", "sortProperty", ":", "params", "[", "'sortProperty'", "]", "=", "sortProperty", "r", "=", "requests", ".", "get", "(", "url", ",", "params", ")", "r", ".", "raise_for_status", "(", ")", "return", "r", ".", "json", "(", ")" ]
Request service locations Returns ------- dict
[ "Request", "service", "locations" ]
train
https://github.com/bertouttier/solaredge/blob/2c94b38527374b0abd088e2819455b332bc5153a/solaredge/solaredge.py#L30-L57
google/dotty
efilter/parsers/common/token_stream.py
TokenStream.match
def match(self, f, *args): """Match grammar function 'f' against next token and set 'self.matched'. Arguments: f: A grammar function - see efilter.parsers.common.grammar. Must return TokenMatch or None. args: Passed to 'f', if any. Returns: Instance of efilter.parsers.common.grammar.TokenMatch or None. Comment: If a match is returned, it will also be stored in self.matched. """ try: match = f(self.tokenizer, *args) except StopIteration: # The grammar function might have tried to access more tokens than # are available. That's not really an error, it just means it didn't # match. return if match is None: return if not isinstance(match, grammar.TokenMatch): raise TypeError("Invalid grammar function %r returned %r." % (f, match)) self.matched = match return match
python
def match(self, f, *args): """Match grammar function 'f' against next token and set 'self.matched'. Arguments: f: A grammar function - see efilter.parsers.common.grammar. Must return TokenMatch or None. args: Passed to 'f', if any. Returns: Instance of efilter.parsers.common.grammar.TokenMatch or None. Comment: If a match is returned, it will also be stored in self.matched. """ try: match = f(self.tokenizer, *args) except StopIteration: # The grammar function might have tried to access more tokens than # are available. That's not really an error, it just means it didn't # match. return if match is None: return if not isinstance(match, grammar.TokenMatch): raise TypeError("Invalid grammar function %r returned %r." % (f, match)) self.matched = match return match
[ "def", "match", "(", "self", ",", "f", ",", "*", "args", ")", ":", "try", ":", "match", "=", "f", "(", "self", ".", "tokenizer", ",", "*", "args", ")", "except", "StopIteration", ":", "# The grammar function might have tried to access more tokens than", "# are available. That's not really an error, it just means it didn't", "# match.", "return", "if", "match", "is", "None", ":", "return", "if", "not", "isinstance", "(", "match", ",", "grammar", ".", "TokenMatch", ")", ":", "raise", "TypeError", "(", "\"Invalid grammar function %r returned %r.\"", "%", "(", "f", ",", "match", ")", ")", "self", ".", "matched", "=", "match", "return", "match" ]
Match grammar function 'f' against next token and set 'self.matched'. Arguments: f: A grammar function - see efilter.parsers.common.grammar. Must return TokenMatch or None. args: Passed to 'f', if any. Returns: Instance of efilter.parsers.common.grammar.TokenMatch or None. Comment: If a match is returned, it will also be stored in self.matched.
[ "Match", "grammar", "function", "f", "against", "next", "token", "and", "set", "self", ".", "matched", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/token_stream.py#L46-L76
google/dotty
efilter/parsers/common/token_stream.py
TokenStream.accept
def accept(self, f, *args): """Like 'match', but consume the token (tokenizer advances.)""" match = self.match(f, *args) if match is None: return self.tokenizer.skip(len(match.tokens)) return match
python
def accept(self, f, *args): """Like 'match', but consume the token (tokenizer advances.)""" match = self.match(f, *args) if match is None: return self.tokenizer.skip(len(match.tokens)) return match
[ "def", "accept", "(", "self", ",", "f", ",", "*", "args", ")", ":", "match", "=", "self", ".", "match", "(", "f", ",", "*", "args", ")", "if", "match", "is", "None", ":", "return", "self", ".", "tokenizer", ".", "skip", "(", "len", "(", "match", ".", "tokens", ")", ")", "return", "match" ]
Like 'match', but consume the token (tokenizer advances.)
[ "Like", "match", "but", "consume", "the", "token", "(", "tokenizer", "advances", ".", ")" ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/token_stream.py#L78-L85
google/dotty
efilter/parsers/common/token_stream.py
TokenStream.reject
def reject(self, f, *args): """Like 'match', but throw a parse error if 'f' matches. This is useful when a parser wants to be strict about specific things being prohibited. For example, DottySQL bans the use of SQL keywords as variable names. """ match = self.match(f, *args) if match: token = self.peek(0) raise errors.EfilterParseError( query=self.tokenizer.source, token=token, message="Was not expecting a %s here." % token.name)
python
def reject(self, f, *args): """Like 'match', but throw a parse error if 'f' matches. This is useful when a parser wants to be strict about specific things being prohibited. For example, DottySQL bans the use of SQL keywords as variable names. """ match = self.match(f, *args) if match: token = self.peek(0) raise errors.EfilterParseError( query=self.tokenizer.source, token=token, message="Was not expecting a %s here." % token.name)
[ "def", "reject", "(", "self", ",", "f", ",", "*", "args", ")", ":", "match", "=", "self", ".", "match", "(", "f", ",", "*", "args", ")", "if", "match", ":", "token", "=", "self", ".", "peek", "(", "0", ")", "raise", "errors", ".", "EfilterParseError", "(", "query", "=", "self", ".", "tokenizer", ".", "source", ",", "token", "=", "token", ",", "message", "=", "\"Was not expecting a %s here.\"", "%", "token", ".", "name", ")" ]
Like 'match', but throw a parse error if 'f' matches. This is useful when a parser wants to be strict about specific things being prohibited. For example, DottySQL bans the use of SQL keywords as variable names.
[ "Like", "match", "but", "throw", "a", "parse", "error", "if", "f", "matches", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/token_stream.py#L87-L99
google/dotty
efilter/parsers/common/token_stream.py
TokenStream.expect
def expect(self, f, *args): """Like 'accept' but throws a parse error if 'f' doesn't match.""" match = self.accept(f, *args) if match: return match try: func_name = f.func_name except AttributeError: func_name = "<unnamed grammar function>" start, end = self.current_position() raise errors.EfilterParseError( query=self.tokenizer.source, start=start, end=end, message="Was expecting %s here." % (func_name))
python
def expect(self, f, *args): """Like 'accept' but throws a parse error if 'f' doesn't match.""" match = self.accept(f, *args) if match: return match try: func_name = f.func_name except AttributeError: func_name = "<unnamed grammar function>" start, end = self.current_position() raise errors.EfilterParseError( query=self.tokenizer.source, start=start, end=end, message="Was expecting %s here." % (func_name))
[ "def", "expect", "(", "self", ",", "f", ",", "*", "args", ")", ":", "match", "=", "self", ".", "accept", "(", "f", ",", "*", "args", ")", "if", "match", ":", "return", "match", "try", ":", "func_name", "=", "f", ".", "func_name", "except", "AttributeError", ":", "func_name", "=", "\"<unnamed grammar function>\"", "start", ",", "end", "=", "self", ".", "current_position", "(", ")", "raise", "errors", ".", "EfilterParseError", "(", "query", "=", "self", ".", "tokenizer", ".", "source", ",", "start", "=", "start", ",", "end", "=", "end", ",", "message", "=", "\"Was expecting %s here.\"", "%", "(", "func_name", ")", ")" ]
Like 'accept' but throws a parse error if 'f' doesn't match.
[ "Like", "accept", "but", "throws", "a", "parse", "error", "if", "f", "doesn", "t", "match", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/token_stream.py#L101-L115
google/dotty
efilter/parsers/common/token_stream.py
TokenStream.current_position
def current_position(self): """Return a tuple of (start, end).""" token = self.tokenizer.peek(0) if token: return token.start, token.end return self.tokenizer.position, self.tokenizer.position + 1
python
def current_position(self): """Return a tuple of (start, end).""" token = self.tokenizer.peek(0) if token: return token.start, token.end return self.tokenizer.position, self.tokenizer.position + 1
[ "def", "current_position", "(", "self", ")", ":", "token", "=", "self", ".", "tokenizer", ".", "peek", "(", "0", ")", "if", "token", ":", "return", "token", ".", "start", ",", "token", ".", "end", "return", "self", ".", "tokenizer", ".", "position", ",", "self", ".", "tokenizer", ".", "position", "+", "1" ]
Return a tuple of (start, end).
[ "Return", "a", "tuple", "of", "(", "start", "end", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/token_stream.py#L117-L123
google/dotty
efilter/parsers/common/ast_transforms.py
ComplementEquivalence
def ComplementEquivalence(*args, **kwargs): """Change x != y to not(x == y).""" return ast.Complement( ast.Equivalence(*args, **kwargs), **kwargs)
python
def ComplementEquivalence(*args, **kwargs): """Change x != y to not(x == y).""" return ast.Complement( ast.Equivalence(*args, **kwargs), **kwargs)
[ "def", "ComplementEquivalence", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "ast", ".", "Complement", "(", "ast", ".", "Equivalence", "(", "*", "args", ",", "*", "*", "kwargs", ")", ",", "*", "*", "kwargs", ")" ]
Change x != y to not(x == y).
[ "Change", "x", "!", "=", "y", "to", "not", "(", "x", "==", "y", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/ast_transforms.py#L43-L46
google/dotty
efilter/parsers/common/ast_transforms.py
ComplementMembership
def ComplementMembership(*args, **kwargs): """Change (x not in y) to not(x in y).""" return ast.Complement( ast.Membership(*args, **kwargs), **kwargs)
python
def ComplementMembership(*args, **kwargs): """Change (x not in y) to not(x in y).""" return ast.Complement( ast.Membership(*args, **kwargs), **kwargs)
[ "def", "ComplementMembership", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "ast", ".", "Complement", "(", "ast", ".", "Membership", "(", "*", "args", ",", "*", "*", "kwargs", ")", ",", "*", "*", "kwargs", ")" ]
Change (x not in y) to not(x in y).
[ "Change", "(", "x", "not", "in", "y", ")", "to", "not", "(", "x", "in", "y", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/ast_transforms.py#L49-L52
google/dotty
efilter/parsers/common/ast_transforms.py
ReverseComplementMembership
def ReverseComplementMembership(x, y, **kwargs): """Change (x doesn't contain y) to not(y in x).""" return ast.Complement( ast.Membership(y, x, **kwargs), **kwargs)
python
def ReverseComplementMembership(x, y, **kwargs): """Change (x doesn't contain y) to not(y in x).""" return ast.Complement( ast.Membership(y, x, **kwargs), **kwargs)
[ "def", "ReverseComplementMembership", "(", "x", ",", "y", ",", "*", "*", "kwargs", ")", ":", "return", "ast", ".", "Complement", "(", "ast", ".", "Membership", "(", "y", ",", "x", ",", "*", "*", "kwargs", ")", ",", "*", "*", "kwargs", ")" ]
Change (x doesn't contain y) to not(y in x).
[ "Change", "(", "x", "doesn", "t", "contain", "y", ")", "to", "not", "(", "y", "in", "x", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/ast_transforms.py#L60-L63
google/dotty
efilter/transforms/solve.py
__solve_for_repeated
def __solve_for_repeated(expr, vars): """Helper: solve 'expr' always returning an IRepeated. If the result of solving 'expr' is a list or a tuple of IStructured objects then treat is as a repeated value of IStructured objects because that's what the called meant to do. This is a convenience helper so users of the API don't have to create IRepeated objects. If the result of solving 'expr' is a scalar then return it as a repeated value of one element. Arguments: expr: Expression to solve. vars: The scope. Returns: IRepeated result of solving 'expr'. A booelan to indicate whether the original was repeating. """ var = solve(expr, vars).value if (var and isinstance(var, (tuple, list)) and protocol.implements(var[0], structured.IStructured)): return repeated.meld(*var), False return var, repeated.isrepeating(var)
python
def __solve_for_repeated(expr, vars): """Helper: solve 'expr' always returning an IRepeated. If the result of solving 'expr' is a list or a tuple of IStructured objects then treat is as a repeated value of IStructured objects because that's what the called meant to do. This is a convenience helper so users of the API don't have to create IRepeated objects. If the result of solving 'expr' is a scalar then return it as a repeated value of one element. Arguments: expr: Expression to solve. vars: The scope. Returns: IRepeated result of solving 'expr'. A booelan to indicate whether the original was repeating. """ var = solve(expr, vars).value if (var and isinstance(var, (tuple, list)) and protocol.implements(var[0], structured.IStructured)): return repeated.meld(*var), False return var, repeated.isrepeating(var)
[ "def", "__solve_for_repeated", "(", "expr", ",", "vars", ")", ":", "var", "=", "solve", "(", "expr", ",", "vars", ")", ".", "value", "if", "(", "var", "and", "isinstance", "(", "var", ",", "(", "tuple", ",", "list", ")", ")", "and", "protocol", ".", "implements", "(", "var", "[", "0", "]", ",", "structured", ".", "IStructured", ")", ")", ":", "return", "repeated", ".", "meld", "(", "*", "var", ")", ",", "False", "return", "var", ",", "repeated", ".", "isrepeating", "(", "var", ")" ]
Helper: solve 'expr' always returning an IRepeated. If the result of solving 'expr' is a list or a tuple of IStructured objects then treat is as a repeated value of IStructured objects because that's what the called meant to do. This is a convenience helper so users of the API don't have to create IRepeated objects. If the result of solving 'expr' is a scalar then return it as a repeated value of one element. Arguments: expr: Expression to solve. vars: The scope. Returns: IRepeated result of solving 'expr'. A booelan to indicate whether the original was repeating.
[ "Helper", ":", "solve", "expr", "always", "returning", "an", "IRepeated", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L84-L108
google/dotty
efilter/transforms/solve.py
__solve_for_scalar
def __solve_for_scalar(expr, vars): """Helper: solve 'expr' always returning a scalar (not IRepeated). If the output of 'expr' is a single value or a single RowTuple with a single column then return the value in that column. Otherwise raise. Arguments: expr: Expression to solve. vars: The scope. Returns: A scalar value (not an IRepeated). Raises: EfilterTypeError if it cannot get a scalar. """ var = solve(expr, vars).value try: scalar = repeated.getvalue(var) except TypeError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Wasn't expecting more than one value here. Got %r." % (var,)) if isinstance(scalar, row_tuple.RowTuple): try: return scalar.get_singleton() except ValueError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Was expecting a scalar value here. Got %r." % (scalar,)) else: return scalar
python
def __solve_for_scalar(expr, vars): """Helper: solve 'expr' always returning a scalar (not IRepeated). If the output of 'expr' is a single value or a single RowTuple with a single column then return the value in that column. Otherwise raise. Arguments: expr: Expression to solve. vars: The scope. Returns: A scalar value (not an IRepeated). Raises: EfilterTypeError if it cannot get a scalar. """ var = solve(expr, vars).value try: scalar = repeated.getvalue(var) except TypeError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Wasn't expecting more than one value here. Got %r." % (var,)) if isinstance(scalar, row_tuple.RowTuple): try: return scalar.get_singleton() except ValueError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Was expecting a scalar value here. Got %r." % (scalar,)) else: return scalar
[ "def", "__solve_for_scalar", "(", "expr", ",", "vars", ")", ":", "var", "=", "solve", "(", "expr", ",", "vars", ")", ".", "value", "try", ":", "scalar", "=", "repeated", ".", "getvalue", "(", "var", ")", "except", "TypeError", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Wasn't expecting more than one value here. Got %r.\"", "%", "(", "var", ",", ")", ")", "if", "isinstance", "(", "scalar", ",", "row_tuple", ".", "RowTuple", ")", ":", "try", ":", "return", "scalar", ".", "get_singleton", "(", ")", "except", "ValueError", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Was expecting a scalar value here. Got %r.\"", "%", "(", "scalar", ",", ")", ")", "else", ":", "return", "scalar" ]
Helper: solve 'expr' always returning a scalar (not IRepeated). If the output of 'expr' is a single value or a single RowTuple with a single column then return the value in that column. Otherwise raise. Arguments: expr: Expression to solve. vars: The scope. Returns: A scalar value (not an IRepeated). Raises: EfilterTypeError if it cannot get a scalar.
[ "Helper", ":", "solve", "expr", "always", "returning", "a", "scalar", "(", "not", "IRepeated", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L111-L145
google/dotty
efilter/transforms/solve.py
__solve_and_destructure_repeated
def __solve_and_destructure_repeated(expr, vars): """Helper: solve 'expr' always returning a list of scalars. If the output of 'expr' is one or more row tuples with only a single column then return a repeated value of values in that column. If there are more than one column per row then raise. This returns a list because there's no point in wrapping the scalars in a repeated value for use internal to the implementing solver. Returns: Two values: - An iterator (not an IRepeated!) of scalars. - A boolean to indicate whether the original value was repeating. Raises: EfilterTypeError if the values don't conform. """ iterable, isrepeating = __solve_for_repeated(expr, vars) if iterable is None: return (), isrepeating if not isrepeating: return [iterable], False values = iter(iterable) try: value = next(values) except StopIteration: return (), True if not isinstance(value, row_tuple.RowTuple): result = [value] # We skip type checking the remaining values because it'd be slow. result.extend(values) return result, True try: result = [value.get_singleton()] for value in values: result.append(value.get_singleton()) return result, True except ValueError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Was expecting exactly one column in %r." % (value,))
python
def __solve_and_destructure_repeated(expr, vars): """Helper: solve 'expr' always returning a list of scalars. If the output of 'expr' is one or more row tuples with only a single column then return a repeated value of values in that column. If there are more than one column per row then raise. This returns a list because there's no point in wrapping the scalars in a repeated value for use internal to the implementing solver. Returns: Two values: - An iterator (not an IRepeated!) of scalars. - A boolean to indicate whether the original value was repeating. Raises: EfilterTypeError if the values don't conform. """ iterable, isrepeating = __solve_for_repeated(expr, vars) if iterable is None: return (), isrepeating if not isrepeating: return [iterable], False values = iter(iterable) try: value = next(values) except StopIteration: return (), True if not isinstance(value, row_tuple.RowTuple): result = [value] # We skip type checking the remaining values because it'd be slow. result.extend(values) return result, True try: result = [value.get_singleton()] for value in values: result.append(value.get_singleton()) return result, True except ValueError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Was expecting exactly one column in %r." % (value,))
[ "def", "__solve_and_destructure_repeated", "(", "expr", ",", "vars", ")", ":", "iterable", ",", "isrepeating", "=", "__solve_for_repeated", "(", "expr", ",", "vars", ")", "if", "iterable", "is", "None", ":", "return", "(", ")", ",", "isrepeating", "if", "not", "isrepeating", ":", "return", "[", "iterable", "]", ",", "False", "values", "=", "iter", "(", "iterable", ")", "try", ":", "value", "=", "next", "(", "values", ")", "except", "StopIteration", ":", "return", "(", ")", ",", "True", "if", "not", "isinstance", "(", "value", ",", "row_tuple", ".", "RowTuple", ")", ":", "result", "=", "[", "value", "]", "# We skip type checking the remaining values because it'd be slow.", "result", ".", "extend", "(", "values", ")", "return", "result", ",", "True", "try", ":", "result", "=", "[", "value", ".", "get_singleton", "(", ")", "]", "for", "value", "in", "values", ":", "result", ".", "append", "(", "value", ".", "get_singleton", "(", ")", ")", "return", "result", ",", "True", "except", "ValueError", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Was expecting exactly one column in %r.\"", "%", "(", "value", ",", ")", ")" ]
Helper: solve 'expr' always returning a list of scalars. If the output of 'expr' is one or more row tuples with only a single column then return a repeated value of values in that column. If there are more than one column per row then raise. This returns a list because there's no point in wrapping the scalars in a repeated value for use internal to the implementing solver. Returns: Two values: - An iterator (not an IRepeated!) of scalars. - A boolean to indicate whether the original value was repeating. Raises: EfilterTypeError if the values don't conform.
[ "Helper", ":", "solve", "expr", "always", "returning", "a", "list", "of", "scalars", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L148-L195
google/dotty
efilter/transforms/solve.py
solve_var
def solve_var(expr, vars): """Returns the value of the var named in the expression.""" try: return Result(structured.resolve(vars, expr.value), ()) except (KeyError, AttributeError) as e: # Raise a better exception for accessing a non-existent member. raise errors.EfilterKeyError(root=expr, key=expr.value, message=e, query=expr.source) except (TypeError, ValueError) as e: # Raise a better exception for what is probably a null pointer error. if vars.locals is None: raise errors.EfilterNoneError( root=expr, query=expr.source, message="Trying to access member %r of a null." % expr.value) else: raise errors.EfilterTypeError( root=expr, query=expr.source, message="%r (vars: %r)" % (e, vars)) except NotImplementedError as e: raise errors.EfilterError( root=expr, query=expr.source, message="Trying to access member %r of an instance of %r." % (expr.value, type(vars)))
python
def solve_var(expr, vars): """Returns the value of the var named in the expression.""" try: return Result(structured.resolve(vars, expr.value), ()) except (KeyError, AttributeError) as e: # Raise a better exception for accessing a non-existent member. raise errors.EfilterKeyError(root=expr, key=expr.value, message=e, query=expr.source) except (TypeError, ValueError) as e: # Raise a better exception for what is probably a null pointer error. if vars.locals is None: raise errors.EfilterNoneError( root=expr, query=expr.source, message="Trying to access member %r of a null." % expr.value) else: raise errors.EfilterTypeError( root=expr, query=expr.source, message="%r (vars: %r)" % (e, vars)) except NotImplementedError as e: raise errors.EfilterError( root=expr, query=expr.source, message="Trying to access member %r of an instance of %r." % (expr.value, type(vars)))
[ "def", "solve_var", "(", "expr", ",", "vars", ")", ":", "try", ":", "return", "Result", "(", "structured", ".", "resolve", "(", "vars", ",", "expr", ".", "value", ")", ",", "(", ")", ")", "except", "(", "KeyError", ",", "AttributeError", ")", "as", "e", ":", "# Raise a better exception for accessing a non-existent member.", "raise", "errors", ".", "EfilterKeyError", "(", "root", "=", "expr", ",", "key", "=", "expr", ".", "value", ",", "message", "=", "e", ",", "query", "=", "expr", ".", "source", ")", "except", "(", "TypeError", ",", "ValueError", ")", "as", "e", ":", "# Raise a better exception for what is probably a null pointer error.", "if", "vars", ".", "locals", "is", "None", ":", "raise", "errors", ".", "EfilterNoneError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Trying to access member %r of a null.\"", "%", "expr", ".", "value", ")", "else", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"%r (vars: %r)\"", "%", "(", "e", ",", "vars", ")", ")", "except", "NotImplementedError", "as", "e", ":", "raise", "errors", ".", "EfilterError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Trying to access member %r of an instance of %r.\"", "%", "(", "expr", ".", "value", ",", "type", "(", "vars", ")", ")", ")" ]
Returns the value of the var named in the expression.
[ "Returns", "the", "value", "of", "the", "var", "named", "in", "the", "expression", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L234-L256
google/dotty
efilter/transforms/solve.py
solve_select
def solve_select(expr, vars): """Use IAssociative.select to get key (rhs) from the data (lhs). This operation supports both scalars and repeated values on the LHS - selecting from a repeated value implies a map-like operation and returns a new repeated value. """ data, _ = __solve_for_repeated(expr.lhs, vars) key = solve(expr.rhs, vars).value try: results = [associative.select(d, key) for d in repeated.getvalues(data)] except (KeyError, AttributeError): # Raise a better exception for accessing a non-existent key. raise errors.EfilterKeyError(root=expr, key=key, query=expr.source) except (TypeError, ValueError): # Raise a better exception for what is probably a null pointer error. if vars.locals is None: raise errors.EfilterNoneError( root=expr, query=expr.source, message="Cannot select key %r from a null." % key) else: raise except NotImplementedError: raise errors.EfilterError( root=expr, query=expr.source, message="Cannot select keys from a non-associative value.") return Result(repeated.meld(*results), ())
python
def solve_select(expr, vars): """Use IAssociative.select to get key (rhs) from the data (lhs). This operation supports both scalars and repeated values on the LHS - selecting from a repeated value implies a map-like operation and returns a new repeated value. """ data, _ = __solve_for_repeated(expr.lhs, vars) key = solve(expr.rhs, vars).value try: results = [associative.select(d, key) for d in repeated.getvalues(data)] except (KeyError, AttributeError): # Raise a better exception for accessing a non-existent key. raise errors.EfilterKeyError(root=expr, key=key, query=expr.source) except (TypeError, ValueError): # Raise a better exception for what is probably a null pointer error. if vars.locals is None: raise errors.EfilterNoneError( root=expr, query=expr.source, message="Cannot select key %r from a null." % key) else: raise except NotImplementedError: raise errors.EfilterError( root=expr, query=expr.source, message="Cannot select keys from a non-associative value.") return Result(repeated.meld(*results), ())
[ "def", "solve_select", "(", "expr", ",", "vars", ")", ":", "data", ",", "_", "=", "__solve_for_repeated", "(", "expr", ".", "lhs", ",", "vars", ")", "key", "=", "solve", "(", "expr", ".", "rhs", ",", "vars", ")", ".", "value", "try", ":", "results", "=", "[", "associative", ".", "select", "(", "d", ",", "key", ")", "for", "d", "in", "repeated", ".", "getvalues", "(", "data", ")", "]", "except", "(", "KeyError", ",", "AttributeError", ")", ":", "# Raise a better exception for accessing a non-existent key.", "raise", "errors", ".", "EfilterKeyError", "(", "root", "=", "expr", ",", "key", "=", "key", ",", "query", "=", "expr", ".", "source", ")", "except", "(", "TypeError", ",", "ValueError", ")", ":", "# Raise a better exception for what is probably a null pointer error.", "if", "vars", ".", "locals", "is", "None", ":", "raise", "errors", ".", "EfilterNoneError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Cannot select key %r from a null.\"", "%", "key", ")", "else", ":", "raise", "except", "NotImplementedError", ":", "raise", "errors", ".", "EfilterError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Cannot select keys from a non-associative value.\"", ")", "return", "Result", "(", "repeated", ".", "meld", "(", "*", "results", ")", ",", "(", ")", ")" ]
Use IAssociative.select to get key (rhs) from the data (lhs). This operation supports both scalars and repeated values on the LHS - selecting from a repeated value implies a map-like operation and returns a new repeated value.
[ "Use", "IAssociative", ".", "select", "to", "get", "key", "(", "rhs", ")", "from", "the", "data", "(", "lhs", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L260-L288
google/dotty
efilter/transforms/solve.py
solve_resolve
def solve_resolve(expr, vars): """Use IStructured.resolve to get member (rhs) from the object (lhs). This operation supports both scalars and repeated values on the LHS - resolving from a repeated value implies a map-like operation and returns a new repeated values. """ objs, _ = __solve_for_repeated(expr.lhs, vars) member = solve(expr.rhs, vars).value try: results = [structured.resolve(o, member) for o in repeated.getvalues(objs)] except (KeyError, AttributeError): # Raise a better exception for the non-existent member. raise errors.EfilterKeyError(root=expr.rhs, key=member, query=expr.source) except (TypeError, ValueError): # Is this a null object error? if vars.locals is None: raise errors.EfilterNoneError( root=expr, query=expr.source, message="Cannot resolve member %r from a null." % member) else: raise except NotImplementedError: raise errors.EfilterError( root=expr, query=expr.source, message="Cannot resolve members from a non-structured value.") return Result(repeated.meld(*results), ())
python
def solve_resolve(expr, vars): """Use IStructured.resolve to get member (rhs) from the object (lhs). This operation supports both scalars and repeated values on the LHS - resolving from a repeated value implies a map-like operation and returns a new repeated values. """ objs, _ = __solve_for_repeated(expr.lhs, vars) member = solve(expr.rhs, vars).value try: results = [structured.resolve(o, member) for o in repeated.getvalues(objs)] except (KeyError, AttributeError): # Raise a better exception for the non-existent member. raise errors.EfilterKeyError(root=expr.rhs, key=member, query=expr.source) except (TypeError, ValueError): # Is this a null object error? if vars.locals is None: raise errors.EfilterNoneError( root=expr, query=expr.source, message="Cannot resolve member %r from a null." % member) else: raise except NotImplementedError: raise errors.EfilterError( root=expr, query=expr.source, message="Cannot resolve members from a non-structured value.") return Result(repeated.meld(*results), ())
[ "def", "solve_resolve", "(", "expr", ",", "vars", ")", ":", "objs", ",", "_", "=", "__solve_for_repeated", "(", "expr", ".", "lhs", ",", "vars", ")", "member", "=", "solve", "(", "expr", ".", "rhs", ",", "vars", ")", ".", "value", "try", ":", "results", "=", "[", "structured", ".", "resolve", "(", "o", ",", "member", ")", "for", "o", "in", "repeated", ".", "getvalues", "(", "objs", ")", "]", "except", "(", "KeyError", ",", "AttributeError", ")", ":", "# Raise a better exception for the non-existent member.", "raise", "errors", ".", "EfilterKeyError", "(", "root", "=", "expr", ".", "rhs", ",", "key", "=", "member", ",", "query", "=", "expr", ".", "source", ")", "except", "(", "TypeError", ",", "ValueError", ")", ":", "# Is this a null object error?", "if", "vars", ".", "locals", "is", "None", ":", "raise", "errors", ".", "EfilterNoneError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Cannot resolve member %r from a null.\"", "%", "member", ")", "else", ":", "raise", "except", "NotImplementedError", ":", "raise", "errors", ".", "EfilterError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Cannot resolve members from a non-structured value.\"", ")", "return", "Result", "(", "repeated", ".", "meld", "(", "*", "results", ")", ",", "(", ")", ")" ]
Use IStructured.resolve to get member (rhs) from the object (lhs). This operation supports both scalars and repeated values on the LHS - resolving from a repeated value implies a map-like operation and returns a new repeated values.
[ "Use", "IStructured", ".", "resolve", "to", "get", "member", "(", "rhs", ")", "from", "the", "object", "(", "lhs", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L292-L322
google/dotty
efilter/transforms/solve.py
solve_apply
def solve_apply(expr, vars): """Returns the result of applying function (lhs) to its arguments (rest). We use IApplicative to apply the function, because that gives the host application an opportunity to compare the function being called against a whitelist. EFILTER will never directly call a function that wasn't provided through a protocol implementation. """ func = __solve_for_scalar(expr.func, vars) args = [] kwargs = {} for arg in expr.args: if isinstance(arg, ast.Pair): if not isinstance(arg.lhs, ast.Var): raise errors.EfilterError( root=arg.lhs, message="Invalid argument name.") kwargs[arg.key.value] = solve(arg.value, vars).value else: args.append(solve(arg, vars).value) result = applicative.apply(func, args, kwargs) return Result(result, ())
python
def solve_apply(expr, vars): """Returns the result of applying function (lhs) to its arguments (rest). We use IApplicative to apply the function, because that gives the host application an opportunity to compare the function being called against a whitelist. EFILTER will never directly call a function that wasn't provided through a protocol implementation. """ func = __solve_for_scalar(expr.func, vars) args = [] kwargs = {} for arg in expr.args: if isinstance(arg, ast.Pair): if not isinstance(arg.lhs, ast.Var): raise errors.EfilterError( root=arg.lhs, message="Invalid argument name.") kwargs[arg.key.value] = solve(arg.value, vars).value else: args.append(solve(arg, vars).value) result = applicative.apply(func, args, kwargs) return Result(result, ())
[ "def", "solve_apply", "(", "expr", ",", "vars", ")", ":", "func", "=", "__solve_for_scalar", "(", "expr", ".", "func", ",", "vars", ")", "args", "=", "[", "]", "kwargs", "=", "{", "}", "for", "arg", "in", "expr", ".", "args", ":", "if", "isinstance", "(", "arg", ",", "ast", ".", "Pair", ")", ":", "if", "not", "isinstance", "(", "arg", ".", "lhs", ",", "ast", ".", "Var", ")", ":", "raise", "errors", ".", "EfilterError", "(", "root", "=", "arg", ".", "lhs", ",", "message", "=", "\"Invalid argument name.\"", ")", "kwargs", "[", "arg", ".", "key", ".", "value", "]", "=", "solve", "(", "arg", ".", "value", ",", "vars", ")", ".", "value", "else", ":", "args", ".", "append", "(", "solve", "(", "arg", ",", "vars", ")", ".", "value", ")", "result", "=", "applicative", ".", "apply", "(", "func", ",", "args", ",", "kwargs", ")", "return", "Result", "(", "result", ",", "(", ")", ")" ]
Returns the result of applying function (lhs) to its arguments (rest). We use IApplicative to apply the function, because that gives the host application an opportunity to compare the function being called against a whitelist. EFILTER will never directly call a function that wasn't provided through a protocol implementation.
[ "Returns", "the", "result", "of", "applying", "function", "(", "lhs", ")", "to", "its", "arguments", "(", "rest", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L326-L350
google/dotty
efilter/transforms/solve.py
solve_bind
def solve_bind(expr, vars): """Build a RowTuple from key/value pairs under the bind. The Bind subtree is arranged as follows: Bind | First KV Pair | | First Key Expression | | First Value Expression | Second KV Pair | | Second Key Expression | | Second Value Expression Etc... As we evaluate the subtree, each subsequent KV pair is evaluated with the all previous bingings already in scope. For example: bind(x: 5, y: x + 5) # Will bind y = 10 because x is already available. """ value_expressions = [] keys = [] for pair in expr.children: keys.append(solve(pair.key, vars).value) value_expressions.append(pair.value) result = row_tuple.RowTuple(ordered_columns=keys) intermediate_scope = scope.ScopeStack(vars, result) for idx, value_expression in enumerate(value_expressions): value = solve(value_expression, intermediate_scope).value # Update the intermediate bindings so as to make earlier bindings # already available to the next child-expression. result[keys[idx]] = value return Result(result, ())
python
def solve_bind(expr, vars): """Build a RowTuple from key/value pairs under the bind. The Bind subtree is arranged as follows: Bind | First KV Pair | | First Key Expression | | First Value Expression | Second KV Pair | | Second Key Expression | | Second Value Expression Etc... As we evaluate the subtree, each subsequent KV pair is evaluated with the all previous bingings already in scope. For example: bind(x: 5, y: x + 5) # Will bind y = 10 because x is already available. """ value_expressions = [] keys = [] for pair in expr.children: keys.append(solve(pair.key, vars).value) value_expressions.append(pair.value) result = row_tuple.RowTuple(ordered_columns=keys) intermediate_scope = scope.ScopeStack(vars, result) for idx, value_expression in enumerate(value_expressions): value = solve(value_expression, intermediate_scope).value # Update the intermediate bindings so as to make earlier bindings # already available to the next child-expression. result[keys[idx]] = value return Result(result, ())
[ "def", "solve_bind", "(", "expr", ",", "vars", ")", ":", "value_expressions", "=", "[", "]", "keys", "=", "[", "]", "for", "pair", "in", "expr", ".", "children", ":", "keys", ".", "append", "(", "solve", "(", "pair", ".", "key", ",", "vars", ")", ".", "value", ")", "value_expressions", ".", "append", "(", "pair", ".", "value", ")", "result", "=", "row_tuple", ".", "RowTuple", "(", "ordered_columns", "=", "keys", ")", "intermediate_scope", "=", "scope", ".", "ScopeStack", "(", "vars", ",", "result", ")", "for", "idx", ",", "value_expression", "in", "enumerate", "(", "value_expressions", ")", ":", "value", "=", "solve", "(", "value_expression", ",", "intermediate_scope", ")", ".", "value", "# Update the intermediate bindings so as to make earlier bindings", "# already available to the next child-expression.", "result", "[", "keys", "[", "idx", "]", "]", "=", "value", "return", "Result", "(", "result", ",", "(", ")", ")" ]
Build a RowTuple from key/value pairs under the bind. The Bind subtree is arranged as follows: Bind | First KV Pair | | First Key Expression | | First Value Expression | Second KV Pair | | Second Key Expression | | Second Value Expression Etc... As we evaluate the subtree, each subsequent KV pair is evaluated with the all previous bingings already in scope. For example: bind(x: 5, y: x + 5) # Will bind y = 10 because x is already available.
[ "Build", "a", "RowTuple", "from", "key", "/", "value", "pairs", "under", "the", "bind", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L354-L388
google/dotty
efilter/transforms/solve.py
solve_repeat
def solve_repeat(expr, vars): """Build a repeated value from subexpressions.""" try: result = repeated.meld(*[solve(x, vars).value for x in expr.children]) return Result(result, ()) except TypeError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="All values in a repeated value must be of the same type.")
python
def solve_repeat(expr, vars): """Build a repeated value from subexpressions.""" try: result = repeated.meld(*[solve(x, vars).value for x in expr.children]) return Result(result, ()) except TypeError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="All values in a repeated value must be of the same type.")
[ "def", "solve_repeat", "(", "expr", ",", "vars", ")", ":", "try", ":", "result", "=", "repeated", ".", "meld", "(", "*", "[", "solve", "(", "x", ",", "vars", ")", ".", "value", "for", "x", "in", "expr", ".", "children", "]", ")", "return", "Result", "(", "result", ",", "(", ")", ")", "except", "TypeError", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"All values in a repeated value must be of the same type.\"", ")" ]
Build a repeated value from subexpressions.
[ "Build", "a", "repeated", "value", "from", "subexpressions", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L392-L400
google/dotty
efilter/transforms/solve.py
solve_tuple
def solve_tuple(expr, vars): """Build a tuple from subexpressions.""" result = tuple(solve(x, vars).value for x in expr.children) return Result(result, ())
python
def solve_tuple(expr, vars): """Build a tuple from subexpressions.""" result = tuple(solve(x, vars).value for x in expr.children) return Result(result, ())
[ "def", "solve_tuple", "(", "expr", ",", "vars", ")", ":", "result", "=", "tuple", "(", "solve", "(", "x", ",", "vars", ")", ".", "value", "for", "x", "in", "expr", ".", "children", ")", "return", "Result", "(", "result", ",", "(", ")", ")" ]
Build a tuple from subexpressions.
[ "Build", "a", "tuple", "from", "subexpressions", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L404-L407
google/dotty
efilter/transforms/solve.py
solve_ifelse
def solve_ifelse(expr, vars): """Evaluate conditions and return the one that matches.""" for condition, result in expr.conditions(): if boolean.asbool(solve(condition, vars).value): return solve(result, vars) return solve(expr.default(), vars)
python
def solve_ifelse(expr, vars): """Evaluate conditions and return the one that matches.""" for condition, result in expr.conditions(): if boolean.asbool(solve(condition, vars).value): return solve(result, vars) return solve(expr.default(), vars)
[ "def", "solve_ifelse", "(", "expr", ",", "vars", ")", ":", "for", "condition", ",", "result", "in", "expr", ".", "conditions", "(", ")", ":", "if", "boolean", ".", "asbool", "(", "solve", "(", "condition", ",", "vars", ")", ".", "value", ")", ":", "return", "solve", "(", "result", ",", "vars", ")", "return", "solve", "(", "expr", ".", "default", "(", ")", ",", "vars", ")" ]
Evaluate conditions and return the one that matches.
[ "Evaluate", "conditions", "and", "return", "the", "one", "that", "matches", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L411-L417
google/dotty
efilter/transforms/solve.py
solve_map
def solve_map(expr, vars): """Solves the map-form, by recursively calling its RHS with new vars. let-forms are binary expressions. The LHS should evaluate to an IAssociative that can be used as new vars with which to solve a new query, of which the RHS is the root. In most cases, the LHS will be a Var (var). Typically, map-forms result from the dotty "dot" (.) operator. For example, the query "User.name" will translate to a map-form with the var "User" on LHS and a var to "name" on the RHS. With top-level vars being something like {"User": {"name": "Bob"}}, the Var on the LHS will evaluate to {"name": "Bob"}, which subdict will then be used on the RHS as new vars, and that whole form will evaluate to "Bob". """ lhs_values, _ = __solve_for_repeated(expr.lhs, vars) def lazy_map(): try: for lhs_value in repeated.getvalues(lhs_values): yield solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value)).value except errors.EfilterNoneError as error: error.root = expr raise return Result(repeated.lazy(lazy_map), ())
python
def solve_map(expr, vars): """Solves the map-form, by recursively calling its RHS with new vars. let-forms are binary expressions. The LHS should evaluate to an IAssociative that can be used as new vars with which to solve a new query, of which the RHS is the root. In most cases, the LHS will be a Var (var). Typically, map-forms result from the dotty "dot" (.) operator. For example, the query "User.name" will translate to a map-form with the var "User" on LHS and a var to "name" on the RHS. With top-level vars being something like {"User": {"name": "Bob"}}, the Var on the LHS will evaluate to {"name": "Bob"}, which subdict will then be used on the RHS as new vars, and that whole form will evaluate to "Bob". """ lhs_values, _ = __solve_for_repeated(expr.lhs, vars) def lazy_map(): try: for lhs_value in repeated.getvalues(lhs_values): yield solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value)).value except errors.EfilterNoneError as error: error.root = expr raise return Result(repeated.lazy(lazy_map), ())
[ "def", "solve_map", "(", "expr", ",", "vars", ")", ":", "lhs_values", ",", "_", "=", "__solve_for_repeated", "(", "expr", ".", "lhs", ",", "vars", ")", "def", "lazy_map", "(", ")", ":", "try", ":", "for", "lhs_value", "in", "repeated", ".", "getvalues", "(", "lhs_values", ")", ":", "yield", "solve", "(", "expr", ".", "rhs", ",", "__nest_scope", "(", "expr", ".", "lhs", ",", "vars", ",", "lhs_value", ")", ")", ".", "value", "except", "errors", ".", "EfilterNoneError", "as", "error", ":", "error", ".", "root", "=", "expr", "raise", "return", "Result", "(", "repeated", ".", "lazy", "(", "lazy_map", ")", ",", "(", ")", ")" ]
Solves the map-form, by recursively calling its RHS with new vars. let-forms are binary expressions. The LHS should evaluate to an IAssociative that can be used as new vars with which to solve a new query, of which the RHS is the root. In most cases, the LHS will be a Var (var). Typically, map-forms result from the dotty "dot" (.) operator. For example, the query "User.name" will translate to a map-form with the var "User" on LHS and a var to "name" on the RHS. With top-level vars being something like {"User": {"name": "Bob"}}, the Var on the LHS will evaluate to {"name": "Bob"}, which subdict will then be used on the RHS as new vars, and that whole form will evaluate to "Bob".
[ "Solves", "the", "map", "-", "form", "by", "recursively", "calling", "its", "RHS", "with", "new", "vars", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L421-L446
google/dotty
efilter/transforms/solve.py
solve_let
def solve_let(expr, vars): """Solves a let-form by calling RHS with nested scope.""" lhs_value = solve(expr.lhs, vars).value if not isinstance(lhs_value, structured.IStructured): raise errors.EfilterTypeError( root=expr.lhs, query=expr.original, message="The LHS of 'let' must evaluate to an IStructured. Got %r." % (lhs_value,)) return solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value))
python
def solve_let(expr, vars): """Solves a let-form by calling RHS with nested scope.""" lhs_value = solve(expr.lhs, vars).value if not isinstance(lhs_value, structured.IStructured): raise errors.EfilterTypeError( root=expr.lhs, query=expr.original, message="The LHS of 'let' must evaluate to an IStructured. Got %r." % (lhs_value,)) return solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value))
[ "def", "solve_let", "(", "expr", ",", "vars", ")", ":", "lhs_value", "=", "solve", "(", "expr", ".", "lhs", ",", "vars", ")", ".", "value", "if", "not", "isinstance", "(", "lhs_value", ",", "structured", ".", "IStructured", ")", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ".", "lhs", ",", "query", "=", "expr", ".", "original", ",", "message", "=", "\"The LHS of 'let' must evaluate to an IStructured. Got %r.\"", "%", "(", "lhs_value", ",", ")", ")", "return", "solve", "(", "expr", ".", "rhs", ",", "__nest_scope", "(", "expr", ".", "lhs", ",", "vars", ",", "lhs_value", ")", ")" ]
Solves a let-form by calling RHS with nested scope.
[ "Solves", "a", "let", "-", "form", "by", "calling", "RHS", "with", "nested", "scope", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L450-L459
google/dotty
efilter/transforms/solve.py
solve_filter
def solve_filter(expr, vars): """Filter values on the LHS by evaluating RHS with each value. Returns any LHS values for which RHS evaluates to a true value. """ lhs_values, _ = __solve_for_repeated(expr.lhs, vars) def lazy_filter(): for lhs_value in repeated.getvalues(lhs_values): if solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value)).value: yield lhs_value return Result(repeated.lazy(lazy_filter), ())
python
def solve_filter(expr, vars): """Filter values on the LHS by evaluating RHS with each value. Returns any LHS values for which RHS evaluates to a true value. """ lhs_values, _ = __solve_for_repeated(expr.lhs, vars) def lazy_filter(): for lhs_value in repeated.getvalues(lhs_values): if solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value)).value: yield lhs_value return Result(repeated.lazy(lazy_filter), ())
[ "def", "solve_filter", "(", "expr", ",", "vars", ")", ":", "lhs_values", ",", "_", "=", "__solve_for_repeated", "(", "expr", ".", "lhs", ",", "vars", ")", "def", "lazy_filter", "(", ")", ":", "for", "lhs_value", "in", "repeated", ".", "getvalues", "(", "lhs_values", ")", ":", "if", "solve", "(", "expr", ".", "rhs", ",", "__nest_scope", "(", "expr", ".", "lhs", ",", "vars", ",", "lhs_value", ")", ")", ".", "value", ":", "yield", "lhs_value", "return", "Result", "(", "repeated", ".", "lazy", "(", "lazy_filter", ")", ",", "(", ")", ")" ]
Filter values on the LHS by evaluating RHS with each value. Returns any LHS values for which RHS evaluates to a true value.
[ "Filter", "values", "on", "the", "LHS", "by", "evaluating", "RHS", "with", "each", "value", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L463-L475
google/dotty
efilter/transforms/solve.py
solve_sort
def solve_sort(expr, vars): """Sort values on the LHS by the value they yield when passed to RHS.""" lhs_values = repeated.getvalues(__solve_for_repeated(expr.lhs, vars)[0]) sort_expression = expr.rhs def _key_func(x): return solve(sort_expression, __nest_scope(expr.lhs, vars, x)).value results = ordered.ordered(lhs_values, key_func=_key_func) return Result(repeated.meld(*results), ())
python
def solve_sort(expr, vars): """Sort values on the LHS by the value they yield when passed to RHS.""" lhs_values = repeated.getvalues(__solve_for_repeated(expr.lhs, vars)[0]) sort_expression = expr.rhs def _key_func(x): return solve(sort_expression, __nest_scope(expr.lhs, vars, x)).value results = ordered.ordered(lhs_values, key_func=_key_func) return Result(repeated.meld(*results), ())
[ "def", "solve_sort", "(", "expr", ",", "vars", ")", ":", "lhs_values", "=", "repeated", ".", "getvalues", "(", "__solve_for_repeated", "(", "expr", ".", "lhs", ",", "vars", ")", "[", "0", "]", ")", "sort_expression", "=", "expr", ".", "rhs", "def", "_key_func", "(", "x", ")", ":", "return", "solve", "(", "sort_expression", ",", "__nest_scope", "(", "expr", ".", "lhs", ",", "vars", ",", "x", ")", ")", ".", "value", "results", "=", "ordered", ".", "ordered", "(", "lhs_values", ",", "key_func", "=", "_key_func", ")", "return", "Result", "(", "repeated", ".", "meld", "(", "*", "results", ")", ",", "(", ")", ")" ]
Sort values on the LHS by the value they yield when passed to RHS.
[ "Sort", "values", "on", "the", "LHS", "by", "the", "value", "they", "yield", "when", "passed", "to", "RHS", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L528-L539
google/dotty
efilter/transforms/solve.py
solve_each
def solve_each(expr, vars): """Return True if RHS evaluates to a true value with each state of LHS. If LHS evaluates to a normal IAssociative object then this is the same as a regular let-form, except the return value is always a boolean. If LHS evaluates to a repeared var (see efilter.protocols.repeated) of IAssociative objects then RHS will be evaluated with each state and True will be returned only if each result is true. """ lhs_values, _ = __solve_for_repeated(expr.lhs, vars) for lhs_value in repeated.getvalues(lhs_values): result = solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value)) if not result.value: # Each is required to return an actual boolean. return result._replace(value=False) return Result(True, ())
python
def solve_each(expr, vars): """Return True if RHS evaluates to a true value with each state of LHS. If LHS evaluates to a normal IAssociative object then this is the same as a regular let-form, except the return value is always a boolean. If LHS evaluates to a repeared var (see efilter.protocols.repeated) of IAssociative objects then RHS will be evaluated with each state and True will be returned only if each result is true. """ lhs_values, _ = __solve_for_repeated(expr.lhs, vars) for lhs_value in repeated.getvalues(lhs_values): result = solve(expr.rhs, __nest_scope(expr.lhs, vars, lhs_value)) if not result.value: # Each is required to return an actual boolean. return result._replace(value=False) return Result(True, ())
[ "def", "solve_each", "(", "expr", ",", "vars", ")", ":", "lhs_values", ",", "_", "=", "__solve_for_repeated", "(", "expr", ".", "lhs", ",", "vars", ")", "for", "lhs_value", "in", "repeated", ".", "getvalues", "(", "lhs_values", ")", ":", "result", "=", "solve", "(", "expr", ".", "rhs", ",", "__nest_scope", "(", "expr", ".", "lhs", ",", "vars", ",", "lhs_value", ")", ")", "if", "not", "result", ".", "value", ":", "# Each is required to return an actual boolean.", "return", "result", ".", "_replace", "(", "value", "=", "False", ")", "return", "Result", "(", "True", ",", "(", ")", ")" ]
Return True if RHS evaluates to a true value with each state of LHS. If LHS evaluates to a normal IAssociative object then this is the same as a regular let-form, except the return value is always a boolean. If LHS evaluates to a repeared var (see efilter.protocols.repeated) of IAssociative objects then RHS will be evaluated with each state and True will be returned only if each result is true.
[ "Return", "True", "if", "RHS", "evaluates", "to", "a", "true", "value", "with", "each", "state", "of", "LHS", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L543-L560
google/dotty
efilter/transforms/solve.py
solve_cast
def solve_cast(expr, vars): """Get cast LHS to RHS.""" lhs = solve(expr.lhs, vars).value t = solve(expr.rhs, vars).value if t is None: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Cannot find type named %r." % expr.rhs.value) if not isinstance(t, type): raise errors.EfilterTypeError( root=expr.rhs, query=expr.source, message="%r is not a type and cannot be used with 'cast'." % (t,)) try: cast_value = t(lhs) except TypeError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Invalid cast %s -> %s." % (type(lhs), t)) return Result(cast_value, ())
python
def solve_cast(expr, vars): """Get cast LHS to RHS.""" lhs = solve(expr.lhs, vars).value t = solve(expr.rhs, vars).value if t is None: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Cannot find type named %r." % expr.rhs.value) if not isinstance(t, type): raise errors.EfilterTypeError( root=expr.rhs, query=expr.source, message="%r is not a type and cannot be used with 'cast'." % (t,)) try: cast_value = t(lhs) except TypeError: raise errors.EfilterTypeError( root=expr, query=expr.source, message="Invalid cast %s -> %s." % (type(lhs), t)) return Result(cast_value, ())
[ "def", "solve_cast", "(", "expr", ",", "vars", ")", ":", "lhs", "=", "solve", "(", "expr", ".", "lhs", ",", "vars", ")", ".", "value", "t", "=", "solve", "(", "expr", ".", "rhs", ",", "vars", ")", ".", "value", "if", "t", "is", "None", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Cannot find type named %r.\"", "%", "expr", ".", "rhs", ".", "value", ")", "if", "not", "isinstance", "(", "t", ",", "type", ")", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ".", "rhs", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"%r is not a type and cannot be used with 'cast'.\"", "%", "(", "t", ",", ")", ")", "try", ":", "cast_value", "=", "t", "(", "lhs", ")", "except", "TypeError", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Invalid cast %s -> %s.\"", "%", "(", "type", "(", "lhs", ")", ",", "t", ")", ")", "return", "Result", "(", "cast_value", ",", "(", ")", ")" ]
Get cast LHS to RHS.
[ "Get", "cast", "LHS", "to", "RHS", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L586-L608
google/dotty
efilter/transforms/solve.py
solve_isinstance
def solve_isinstance(expr, vars): """Typecheck whether LHS is type on the RHS.""" lhs = solve(expr.lhs, vars) try: t = solve(expr.rhs, vars).value except errors.EfilterKeyError: t = None if t is None: raise errors.EfilterTypeError( root=expr.rhs, query=expr.source, message="Cannot find type named %r." % expr.rhs.value) if not isinstance(t, type): raise errors.EfilterTypeError( root=expr.rhs, query=expr.source, message="%r is not a type and cannot be used with 'isa'." % (t,)) return Result(protocol.implements(lhs.value, t), ())
python
def solve_isinstance(expr, vars): """Typecheck whether LHS is type on the RHS.""" lhs = solve(expr.lhs, vars) try: t = solve(expr.rhs, vars).value except errors.EfilterKeyError: t = None if t is None: raise errors.EfilterTypeError( root=expr.rhs, query=expr.source, message="Cannot find type named %r." % expr.rhs.value) if not isinstance(t, type): raise errors.EfilterTypeError( root=expr.rhs, query=expr.source, message="%r is not a type and cannot be used with 'isa'." % (t,)) return Result(protocol.implements(lhs.value, t), ())
[ "def", "solve_isinstance", "(", "expr", ",", "vars", ")", ":", "lhs", "=", "solve", "(", "expr", ".", "lhs", ",", "vars", ")", "try", ":", "t", "=", "solve", "(", "expr", ".", "rhs", ",", "vars", ")", ".", "value", "except", "errors", ".", "EfilterKeyError", ":", "t", "=", "None", "if", "t", "is", "None", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ".", "rhs", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"Cannot find type named %r.\"", "%", "expr", ".", "rhs", ".", "value", ")", "if", "not", "isinstance", "(", "t", ",", "type", ")", ":", "raise", "errors", ".", "EfilterTypeError", "(", "root", "=", "expr", ".", "rhs", ",", "query", "=", "expr", ".", "source", ",", "message", "=", "\"%r is not a type and cannot be used with 'isa'.\"", "%", "(", "t", ",", ")", ")", "return", "Result", "(", "protocol", ".", "implements", "(", "lhs", ".", "value", ",", "t", ")", ",", "(", ")", ")" ]
Typecheck whether LHS is type on the RHS.
[ "Typecheck", "whether", "LHS", "is", "type", "on", "the", "RHS", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/solve.py#L612-L631
radical-cybertools/radical.entk
setup.py
set_version
def set_version(mod_root): """ mod_root a VERSION file containes the version strings is created in mod_root, during installation. That file is used at runtime to get the version information. """ try: version_base = None version_detail = None # get version from './VERSION' src_root = os.path.dirname(__file__) if not src_root: src_root = '.' with open(src_root + '/VERSION', 'r') as f: version_base = f.readline().strip() # attempt to get version detail information from git # We only do that though if we are in a repo root dir, # ie. if 'git rev-parse --show-prefix' returns an empty string -- # otherwise we get confused if the ve lives beneath another repository, # and the pip version used uses an install tmp dir in the ve space # instead of /tmp (which seems to happen with some pip/setuptools # versions). p = sp.Popen('cd %s ; ' 'test -z `git rev-parse --show-prefix` || exit -1; ' 'tag=`git describe --tags --always` 2>/dev/null ; ' 'branch=`git branch | grep -e "^*" | cut -f 2- -d " "` 2>/dev/null ; ' 'echo $tag@$branch' % src_root, stdout=sp.PIPE, stderr=sp.STDOUT, shell=True) version_detail = str(p.communicate()[0].strip()) version_detail = version_detail.replace('detached from ', 'detached-') # remove all non-alphanumeric (and then some) chars version_detail = re.sub('[/ ]+', '-', version_detail) version_detail = re.sub('[^a-zA-Z0-9_+@.-]+', '', version_detail) if p.returncode != 0 or \ version_detail == '@' or \ 'git-error' in version_detail or \ 'not-a-git-repo' in version_detail or \ 'not-found' in version_detail or \ 'fatal' in version_detail : version = version_base elif '@' not in version_base: version = '%s-%s' % (version_base, version_detail) else: version = version_base # make sure the version files exist for the runtime version inspection path = '%s/%s' % (src_root, mod_root) with open(path + "/VERSION", "w") as f: f.write(version + "\n") sdist_name = "%s-%s.tar.gz" % (name, version) sdist_name = sdist_name.replace('/', '-') sdist_name = sdist_name.replace('@', '-') sdist_name = sdist_name.replace('#', '-') sdist_name = sdist_name.replace('_', '-') if '--record' in sys.argv or \ 'bdist_egg' in sys.argv or \ 'bdist_wheel' in sys.argv : # pip install stage 2 or easy_install stage 1 # # pip install will untar the sdist in a tmp tree. In that tmp # tree, we won't be able to derive git version tags -- so we pack the # formerly derived version as ./VERSION shutil.move("VERSION", "VERSION.bak") # backup version shutil.copy("%s/VERSION" % path, "VERSION") # use full version instead os.system ("python setup.py sdist") # build sdist shutil.copy('dist/%s' % sdist_name, '%s/%s' % (mod_root, sdist_name)) # copy into tree shutil.move("VERSION.bak", "VERSION") # restore version with open(path + "/SDIST", "w") as f: f.write(sdist_name + "\n") return version_base, version_detail, sdist_name except Exception as e : raise RuntimeError('Could not extract/set version: %s' % e)
python
def set_version(mod_root): """ mod_root a VERSION file containes the version strings is created in mod_root, during installation. That file is used at runtime to get the version information. """ try: version_base = None version_detail = None # get version from './VERSION' src_root = os.path.dirname(__file__) if not src_root: src_root = '.' with open(src_root + '/VERSION', 'r') as f: version_base = f.readline().strip() # attempt to get version detail information from git # We only do that though if we are in a repo root dir, # ie. if 'git rev-parse --show-prefix' returns an empty string -- # otherwise we get confused if the ve lives beneath another repository, # and the pip version used uses an install tmp dir in the ve space # instead of /tmp (which seems to happen with some pip/setuptools # versions). p = sp.Popen('cd %s ; ' 'test -z `git rev-parse --show-prefix` || exit -1; ' 'tag=`git describe --tags --always` 2>/dev/null ; ' 'branch=`git branch | grep -e "^*" | cut -f 2- -d " "` 2>/dev/null ; ' 'echo $tag@$branch' % src_root, stdout=sp.PIPE, stderr=sp.STDOUT, shell=True) version_detail = str(p.communicate()[0].strip()) version_detail = version_detail.replace('detached from ', 'detached-') # remove all non-alphanumeric (and then some) chars version_detail = re.sub('[/ ]+', '-', version_detail) version_detail = re.sub('[^a-zA-Z0-9_+@.-]+', '', version_detail) if p.returncode != 0 or \ version_detail == '@' or \ 'git-error' in version_detail or \ 'not-a-git-repo' in version_detail or \ 'not-found' in version_detail or \ 'fatal' in version_detail : version = version_base elif '@' not in version_base: version = '%s-%s' % (version_base, version_detail) else: version = version_base # make sure the version files exist for the runtime version inspection path = '%s/%s' % (src_root, mod_root) with open(path + "/VERSION", "w") as f: f.write(version + "\n") sdist_name = "%s-%s.tar.gz" % (name, version) sdist_name = sdist_name.replace('/', '-') sdist_name = sdist_name.replace('@', '-') sdist_name = sdist_name.replace('#', '-') sdist_name = sdist_name.replace('_', '-') if '--record' in sys.argv or \ 'bdist_egg' in sys.argv or \ 'bdist_wheel' in sys.argv : # pip install stage 2 or easy_install stage 1 # # pip install will untar the sdist in a tmp tree. In that tmp # tree, we won't be able to derive git version tags -- so we pack the # formerly derived version as ./VERSION shutil.move("VERSION", "VERSION.bak") # backup version shutil.copy("%s/VERSION" % path, "VERSION") # use full version instead os.system ("python setup.py sdist") # build sdist shutil.copy('dist/%s' % sdist_name, '%s/%s' % (mod_root, sdist_name)) # copy into tree shutil.move("VERSION.bak", "VERSION") # restore version with open(path + "/SDIST", "w") as f: f.write(sdist_name + "\n") return version_base, version_detail, sdist_name except Exception as e : raise RuntimeError('Could not extract/set version: %s' % e)
[ "def", "set_version", "(", "mod_root", ")", ":", "try", ":", "version_base", "=", "None", "version_detail", "=", "None", "# get version from './VERSION'", "src_root", "=", "os", ".", "path", ".", "dirname", "(", "__file__", ")", "if", "not", "src_root", ":", "src_root", "=", "'.'", "with", "open", "(", "src_root", "+", "'/VERSION'", ",", "'r'", ")", "as", "f", ":", "version_base", "=", "f", ".", "readline", "(", ")", ".", "strip", "(", ")", "# attempt to get version detail information from git", "# We only do that though if we are in a repo root dir,", "# ie. if 'git rev-parse --show-prefix' returns an empty string --", "# otherwise we get confused if the ve lives beneath another repository,", "# and the pip version used uses an install tmp dir in the ve space", "# instead of /tmp (which seems to happen with some pip/setuptools", "# versions).", "p", "=", "sp", ".", "Popen", "(", "'cd %s ; '", "'test -z `git rev-parse --show-prefix` || exit -1; '", "'tag=`git describe --tags --always` 2>/dev/null ; '", "'branch=`git branch | grep -e \"^*\" | cut -f 2- -d \" \"` 2>/dev/null ; '", "'echo $tag@$branch'", "%", "src_root", ",", "stdout", "=", "sp", ".", "PIPE", ",", "stderr", "=", "sp", ".", "STDOUT", ",", "shell", "=", "True", ")", "version_detail", "=", "str", "(", "p", ".", "communicate", "(", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "version_detail", "=", "version_detail", ".", "replace", "(", "'detached from '", ",", "'detached-'", ")", "# remove all non-alphanumeric (and then some) chars", "version_detail", "=", "re", ".", "sub", "(", "'[/ ]+'", ",", "'-'", ",", "version_detail", ")", "version_detail", "=", "re", ".", "sub", "(", "'[^a-zA-Z0-9_+@.-]+'", ",", "''", ",", "version_detail", ")", "if", "p", ".", "returncode", "!=", "0", "or", "version_detail", "==", "'@'", "or", "'git-error'", "in", "version_detail", "or", "'not-a-git-repo'", "in", "version_detail", "or", "'not-found'", "in", "version_detail", "or", "'fatal'", "in", "version_detail", ":", "version", "=", "version_base", "elif", "'@'", "not", "in", "version_base", ":", "version", "=", "'%s-%s'", "%", "(", "version_base", ",", "version_detail", ")", "else", ":", "version", "=", "version_base", "# make sure the version files exist for the runtime version inspection", "path", "=", "'%s/%s'", "%", "(", "src_root", ",", "mod_root", ")", "with", "open", "(", "path", "+", "\"/VERSION\"", ",", "\"w\"", ")", "as", "f", ":", "f", ".", "write", "(", "version", "+", "\"\\n\"", ")", "sdist_name", "=", "\"%s-%s.tar.gz\"", "%", "(", "name", ",", "version", ")", "sdist_name", "=", "sdist_name", ".", "replace", "(", "'/'", ",", "'-'", ")", "sdist_name", "=", "sdist_name", ".", "replace", "(", "'@'", ",", "'-'", ")", "sdist_name", "=", "sdist_name", ".", "replace", "(", "'#'", ",", "'-'", ")", "sdist_name", "=", "sdist_name", ".", "replace", "(", "'_'", ",", "'-'", ")", "if", "'--record'", "in", "sys", ".", "argv", "or", "'bdist_egg'", "in", "sys", ".", "argv", "or", "'bdist_wheel'", "in", "sys", ".", "argv", ":", "# pip install stage 2 or easy_install stage 1", "#", "# pip install will untar the sdist in a tmp tree. In that tmp", "# tree, we won't be able to derive git version tags -- so we pack the", "# formerly derived version as ./VERSION", "shutil", ".", "move", "(", "\"VERSION\"", ",", "\"VERSION.bak\"", ")", "# backup version", "shutil", ".", "copy", "(", "\"%s/VERSION\"", "%", "path", ",", "\"VERSION\"", ")", "# use full version instead", "os", ".", "system", "(", "\"python setup.py sdist\"", ")", "# build sdist", "shutil", ".", "copy", "(", "'dist/%s'", "%", "sdist_name", ",", "'%s/%s'", "%", "(", "mod_root", ",", "sdist_name", ")", ")", "# copy into tree", "shutil", ".", "move", "(", "\"VERSION.bak\"", ",", "\"VERSION\"", ")", "# restore version", "with", "open", "(", "path", "+", "\"/SDIST\"", ",", "\"w\"", ")", "as", "f", ":", "f", ".", "write", "(", "sdist_name", "+", "\"\\n\"", ")", "return", "version_base", ",", "version_detail", ",", "sdist_name", "except", "Exception", "as", "e", ":", "raise", "RuntimeError", "(", "'Could not extract/set version: %s'", "%", "e", ")" ]
mod_root a VERSION file containes the version strings is created in mod_root, during installation. That file is used at runtime to get the version information.
[ "mod_root", "a", "VERSION", "file", "containes", "the", "version", "strings", "is", "created", "in", "mod_root", "during", "installation", ".", "That", "file", "is", "used", "at", "runtime", "to", "get", "the", "version", "information", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/setup.py#L26-L111
radical-cybertools/radical.entk
setup.py
makeDataFiles
def makeDataFiles(prefix, dir): """ Create distutils data_files structure from dir distutil will copy all file rooted under dir into prefix, excluding dir itself, just like 'ditto src dst' works, and unlike 'cp -r src dst, which copy src into dst'. Typical usage: # install the contents of 'wiki' under sys.prefix+'share/moin' data_files = makeDataFiles('share/moin', 'wiki') For this directory structure: root file1 file2 dir file subdir file makeDataFiles('prefix', 'root') will create this distutil data_files structure: [('prefix', ['file1', 'file2']), ('prefix/dir', ['file']), ('prefix/dir/subdir', ['file'])] """ # Strip 'dir/' from of path before joining with prefix dir = dir.rstrip('/') strip = len(dir) + 1 found = [] os.path.walk(dir, visit, (prefix, strip, found)) #print found[0] return found[0]
python
def makeDataFiles(prefix, dir): """ Create distutils data_files structure from dir distutil will copy all file rooted under dir into prefix, excluding dir itself, just like 'ditto src dst' works, and unlike 'cp -r src dst, which copy src into dst'. Typical usage: # install the contents of 'wiki' under sys.prefix+'share/moin' data_files = makeDataFiles('share/moin', 'wiki') For this directory structure: root file1 file2 dir file subdir file makeDataFiles('prefix', 'root') will create this distutil data_files structure: [('prefix', ['file1', 'file2']), ('prefix/dir', ['file']), ('prefix/dir/subdir', ['file'])] """ # Strip 'dir/' from of path before joining with prefix dir = dir.rstrip('/') strip = len(dir) + 1 found = [] os.path.walk(dir, visit, (prefix, strip, found)) #print found[0] return found[0]
[ "def", "makeDataFiles", "(", "prefix", ",", "dir", ")", ":", "# Strip 'dir/' from of path before joining with prefix", "dir", "=", "dir", ".", "rstrip", "(", "'/'", ")", "strip", "=", "len", "(", "dir", ")", "+", "1", "found", "=", "[", "]", "os", ".", "path", ".", "walk", "(", "dir", ",", "visit", ",", "(", "prefix", ",", "strip", ",", "found", ")", ")", "#print found[0]", "return", "found", "[", "0", "]" ]
Create distutils data_files structure from dir distutil will copy all file rooted under dir into prefix, excluding dir itself, just like 'ditto src dst' works, and unlike 'cp -r src dst, which copy src into dst'. Typical usage: # install the contents of 'wiki' under sys.prefix+'share/moin' data_files = makeDataFiles('share/moin', 'wiki') For this directory structure: root file1 file2 dir file subdir file makeDataFiles('prefix', 'root') will create this distutil data_files structure: [('prefix', ['file1', 'file2']), ('prefix/dir', ['file']), ('prefix/dir/subdir', ['file'])]
[ "Create", "distutils", "data_files", "structure", "from", "dir", "distutil", "will", "copy", "all", "file", "rooted", "under", "dir", "into", "prefix", "excluding", "dir", "itself", "just", "like", "ditto", "src", "dst", "works", "and", "unlike", "cp", "-", "r", "src", "dst", "which", "copy", "src", "into", "dst", ".", "Typical", "usage", ":", "#", "install", "the", "contents", "of", "wiki", "under", "sys", ".", "prefix", "+", "share", "/", "moin", "data_files", "=", "makeDataFiles", "(", "share", "/", "moin", "wiki", ")", "For", "this", "directory", "structure", ":", "root", "file1", "file2", "dir", "file", "subdir", "file", "makeDataFiles", "(", "prefix", "root", ")", "will", "create", "this", "distutil", "data_files", "structure", ":", "[", "(", "prefix", "[", "file1", "file2", "]", ")", "(", "prefix", "/", "dir", "[", "file", "]", ")", "(", "prefix", "/", "dir", "/", "subdir", "[", "file", "]", ")", "]" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/setup.py#L126-L153
radical-cybertools/radical.entk
setup.py
visit
def visit((prefix, strip, found), dirname, names): """ Visit directory, create distutil tuple Add distutil tuple for each directory using this format: (destination, [dirname/file1, dirname/file2, ...]) distutil will copy later file1, file2, ... info destination. """ files = [] # Iterate over a copy of names, modify names for name in names[:]: path = os.path.join(dirname, name) # Ignore directories - we will visit later if os.path.isdir(path): # Remove directories we don't want to visit later if isbad(name): names.remove(name) continue elif isgood(name): files.append(path) destination = os.path.join(prefix, dirname[strip:]) found.append((destination, files))
python
def visit((prefix, strip, found), dirname, names): """ Visit directory, create distutil tuple Add distutil tuple for each directory using this format: (destination, [dirname/file1, dirname/file2, ...]) distutil will copy later file1, file2, ... info destination. """ files = [] # Iterate over a copy of names, modify names for name in names[:]: path = os.path.join(dirname, name) # Ignore directories - we will visit later if os.path.isdir(path): # Remove directories we don't want to visit later if isbad(name): names.remove(name) continue elif isgood(name): files.append(path) destination = os.path.join(prefix, dirname[strip:]) found.append((destination, files))
[ "def", "visit", "(", "(", "prefix", ",", "strip", ",", "found", ")", ",", "dirname", ",", "names", ")", ":", "files", "=", "[", "]", "# Iterate over a copy of names, modify names", "for", "name", "in", "names", "[", ":", "]", ":", "path", "=", "os", ".", "path", ".", "join", "(", "dirname", ",", "name", ")", "# Ignore directories - we will visit later", "if", "os", ".", "path", ".", "isdir", "(", "path", ")", ":", "# Remove directories we don't want to visit later", "if", "isbad", "(", "name", ")", ":", "names", ".", "remove", "(", "name", ")", "continue", "elif", "isgood", "(", "name", ")", ":", "files", ".", "append", "(", "path", ")", "destination", "=", "os", ".", "path", ".", "join", "(", "prefix", ",", "dirname", "[", "strip", ":", "]", ")", "found", ".", "append", "(", "(", "destination", ",", "files", ")", ")" ]
Visit directory, create distutil tuple Add distutil tuple for each directory using this format: (destination, [dirname/file1, dirname/file2, ...]) distutil will copy later file1, file2, ... info destination.
[ "Visit", "directory", "create", "distutil", "tuple", "Add", "distutil", "tuple", "for", "each", "directory", "using", "this", "format", ":", "(", "destination", "[", "dirname", "/", "file1", "dirname", "/", "file2", "...", "]", ")", "distutil", "will", "copy", "later", "file1", "file2", "...", "info", "destination", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/setup.py#L155-L174
radical-cybertools/radical.entk
setup.py
isgood
def isgood(name): """ Whether name should be installed """ if not isbad(name): if name.endswith('.py') or name.endswith('.json') or name.endswith('.tar'): return True return False
python
def isgood(name): """ Whether name should be installed """ if not isbad(name): if name.endswith('.py') or name.endswith('.json') or name.endswith('.tar'): return True return False
[ "def", "isgood", "(", "name", ")", ":", "if", "not", "isbad", "(", "name", ")", ":", "if", "name", ".", "endswith", "(", "'.py'", ")", "or", "name", ".", "endswith", "(", "'.json'", ")", "or", "name", ".", "endswith", "(", "'.tar'", ")", ":", "return", "True", "return", "False" ]
Whether name should be installed
[ "Whether", "name", "should", "be", "installed" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/setup.py#L183-L188
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor._initialize_workflow
def _initialize_workflow(self): """ **Purpose**: Initialize the PST of the workflow with a uid and type checks """ try: self._prof.prof('initializing workflow', uid=self._uid) for p in self._workflow: p._assign_uid(self._sid) self._prof.prof('workflow initialized', uid=self._uid) except Exception, ex: self._logger.exception( 'Fatal error while initializing workflow: %s' % ex) raise
python
def _initialize_workflow(self): """ **Purpose**: Initialize the PST of the workflow with a uid and type checks """ try: self._prof.prof('initializing workflow', uid=self._uid) for p in self._workflow: p._assign_uid(self._sid) self._prof.prof('workflow initialized', uid=self._uid) except Exception, ex: self._logger.exception( 'Fatal error while initializing workflow: %s' % ex) raise
[ "def", "_initialize_workflow", "(", "self", ")", ":", "try", ":", "self", ".", "_prof", ".", "prof", "(", "'initializing workflow'", ",", "uid", "=", "self", ".", "_uid", ")", "for", "p", "in", "self", ".", "_workflow", ":", "p", ".", "_assign_uid", "(", "self", ".", "_sid", ")", "self", ".", "_prof", ".", "prof", "(", "'workflow initialized'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Fatal error while initializing workflow: %s'", "%", "ex", ")", "raise" ]
**Purpose**: Initialize the PST of the workflow with a uid and type checks
[ "**", "Purpose", "**", ":", "Initialize", "the", "PST", "of", "the", "workflow", "with", "a", "uid", "and", "type", "checks" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L87-L104
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor._enqueue
def _enqueue(self, local_prof): """ **Purpose**: This is the function that is run in the enqueue thread. This function extracts Tasks from the copy of workflow that exists in the WFprocessor object and pushes them to the queues in the pending_q list. Since this thread works on the copy of the workflow, every state update to the Task, Stage and Pipeline is communicated back to the AppManager (master process) via the 'sync_with_master' function that has dedicated queues to communicate with the master. Details: Termination condition of this thread is set by the wfp process. """ try: local_prof.prof('enqueue-thread started', uid=self._uid) self._logger.info('enqueue-thread started') # Acquire a connection+channel to the rmq server mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() last = time.time() while not self._enqueue_thread_terminate.is_set(): ''' We iterate through all pipelines to collect tasks from stages that are pending scheduling. Once collected, these tasks will be communicated to the tmgr in bulk. ''' workload = [] scheduled_stages = [] for pipe in self._workflow: with pipe.lock: if ((not pipe.completed) and (not pipe.state == states.SUSPENDED)): # Test if the pipeline is already in the final state if pipe.state in states.FINAL: continue elif pipe.state == states.INITIAL: # Set state of pipeline to SCHEDULING if it is in INITIAL transition(obj=pipe, obj_type='Pipeline', new_state=states.SCHEDULING, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) executable_stage = pipe.stages[pipe.current_stage - 1] if not executable_stage.uid: executable_stage.parent_pipeline['uid'] = pipe.uid executable_stage.parent_pipeline['name'] = pipe.name executable_stage._assign_uid(self._sid) if executable_stage.state in [states.INITIAL, states.SCHEDULED]: if executable_stage.state == states.INITIAL: transition(obj=executable_stage, obj_type='Stage', new_state=states.SCHEDULING, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) executable_tasks = executable_stage.tasks for executable_task in executable_tasks: if (executable_task.state == states.INITIAL)or \ ((executable_task.state == states.FAILED)and(self._resubmit_failed)): # Set state of Tasks in current Stage to SCHEDULING transition(obj=executable_task, obj_type='Task', new_state=states.SCHEDULING, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # task_as_dict = json.dumps(executable_task.to_dict()) workload.append(executable_task) if executable_stage not in scheduled_stages: scheduled_stages.append( executable_stage) if workload: # Put the task on one of the pending_queues workload_as_dict = list() for task in workload: workload_as_dict.append(task.to_dict()) workload_as_dict = json.dumps(workload_as_dict) mq_channel.basic_publish(exchange='', routing_key=self._pending_queue[0], body=workload_as_dict # properties=pika.BasicProperties( # make message persistent # delivery_mode = 2) ) for task in workload: # Set state of Tasks in current Stage to SCHEDULED transition(obj=task, obj_type='Task', new_state=states.SCHEDULED, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) self._logger.debug( 'Task %s published to pending queue' % task.uid) if scheduled_stages: for executable_stage in scheduled_stages: transition(obj=executable_stage, obj_type='Stage', new_state=states.SCHEDULED, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Appease pika cos it thinks the connection is dead now = time.time() if now - last >= self._rmq_ping_interval: mq_connection.process_data_events() last = now self._logger.info('Enqueue thread terminated') mq_connection.close() local_prof.prof('terminating enqueue-thread', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel enqueuer thread gracefully...') mq_connection.close() raise KeyboardInterrupt except Exception, ex: self._logger.exception('Error in enqueue-thread: %s' % ex) try: mq_connection.close() except Exception as ex: self._logger.warning('mq_connection not created, %s' % ex) raise
python
def _enqueue(self, local_prof): """ **Purpose**: This is the function that is run in the enqueue thread. This function extracts Tasks from the copy of workflow that exists in the WFprocessor object and pushes them to the queues in the pending_q list. Since this thread works on the copy of the workflow, every state update to the Task, Stage and Pipeline is communicated back to the AppManager (master process) via the 'sync_with_master' function that has dedicated queues to communicate with the master. Details: Termination condition of this thread is set by the wfp process. """ try: local_prof.prof('enqueue-thread started', uid=self._uid) self._logger.info('enqueue-thread started') # Acquire a connection+channel to the rmq server mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() last = time.time() while not self._enqueue_thread_terminate.is_set(): ''' We iterate through all pipelines to collect tasks from stages that are pending scheduling. Once collected, these tasks will be communicated to the tmgr in bulk. ''' workload = [] scheduled_stages = [] for pipe in self._workflow: with pipe.lock: if ((not pipe.completed) and (not pipe.state == states.SUSPENDED)): # Test if the pipeline is already in the final state if pipe.state in states.FINAL: continue elif pipe.state == states.INITIAL: # Set state of pipeline to SCHEDULING if it is in INITIAL transition(obj=pipe, obj_type='Pipeline', new_state=states.SCHEDULING, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) executable_stage = pipe.stages[pipe.current_stage - 1] if not executable_stage.uid: executable_stage.parent_pipeline['uid'] = pipe.uid executable_stage.parent_pipeline['name'] = pipe.name executable_stage._assign_uid(self._sid) if executable_stage.state in [states.INITIAL, states.SCHEDULED]: if executable_stage.state == states.INITIAL: transition(obj=executable_stage, obj_type='Stage', new_state=states.SCHEDULING, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) executable_tasks = executable_stage.tasks for executable_task in executable_tasks: if (executable_task.state == states.INITIAL)or \ ((executable_task.state == states.FAILED)and(self._resubmit_failed)): # Set state of Tasks in current Stage to SCHEDULING transition(obj=executable_task, obj_type='Task', new_state=states.SCHEDULING, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # task_as_dict = json.dumps(executable_task.to_dict()) workload.append(executable_task) if executable_stage not in scheduled_stages: scheduled_stages.append( executable_stage) if workload: # Put the task on one of the pending_queues workload_as_dict = list() for task in workload: workload_as_dict.append(task.to_dict()) workload_as_dict = json.dumps(workload_as_dict) mq_channel.basic_publish(exchange='', routing_key=self._pending_queue[0], body=workload_as_dict # properties=pika.BasicProperties( # make message persistent # delivery_mode = 2) ) for task in workload: # Set state of Tasks in current Stage to SCHEDULED transition(obj=task, obj_type='Task', new_state=states.SCHEDULED, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) self._logger.debug( 'Task %s published to pending queue' % task.uid) if scheduled_stages: for executable_stage in scheduled_stages: transition(obj=executable_stage, obj_type='Stage', new_state=states.SCHEDULED, channel=mq_channel, queue='%s-enq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Appease pika cos it thinks the connection is dead now = time.time() if now - last >= self._rmq_ping_interval: mq_connection.process_data_events() last = now self._logger.info('Enqueue thread terminated') mq_connection.close() local_prof.prof('terminating enqueue-thread', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel enqueuer thread gracefully...') mq_connection.close() raise KeyboardInterrupt except Exception, ex: self._logger.exception('Error in enqueue-thread: %s' % ex) try: mq_connection.close() except Exception as ex: self._logger.warning('mq_connection not created, %s' % ex) raise
[ "def", "_enqueue", "(", "self", ",", "local_prof", ")", ":", "try", ":", "local_prof", ".", "prof", "(", "'enqueue-thread started'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'enqueue-thread started'", ")", "# Acquire a connection+channel to the rmq server", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "self", ".", "_mq_hostname", ",", "port", "=", "self", ".", "_port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "last", "=", "time", ".", "time", "(", ")", "while", "not", "self", ".", "_enqueue_thread_terminate", ".", "is_set", "(", ")", ":", "'''\n We iterate through all pipelines to collect tasks from\n stages that are pending scheduling. Once collected, these tasks\n will be communicated to the tmgr in bulk.\n '''", "workload", "=", "[", "]", "scheduled_stages", "=", "[", "]", "for", "pipe", "in", "self", ".", "_workflow", ":", "with", "pipe", ".", "lock", ":", "if", "(", "(", "not", "pipe", ".", "completed", ")", "and", "(", "not", "pipe", ".", "state", "==", "states", ".", "SUSPENDED", ")", ")", ":", "# Test if the pipeline is already in the final state", "if", "pipe", ".", "state", "in", "states", ".", "FINAL", ":", "continue", "elif", "pipe", ".", "state", "==", "states", ".", "INITIAL", ":", "# Set state of pipeline to SCHEDULING if it is in INITIAL", "transition", "(", "obj", "=", "pipe", ",", "obj_type", "=", "'Pipeline'", ",", "new_state", "=", "states", ".", "SCHEDULING", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "executable_stage", "=", "pipe", ".", "stages", "[", "pipe", ".", "current_stage", "-", "1", "]", "if", "not", "executable_stage", ".", "uid", ":", "executable_stage", ".", "parent_pipeline", "[", "'uid'", "]", "=", "pipe", ".", "uid", "executable_stage", ".", "parent_pipeline", "[", "'name'", "]", "=", "pipe", ".", "name", "executable_stage", ".", "_assign_uid", "(", "self", ".", "_sid", ")", "if", "executable_stage", ".", "state", "in", "[", "states", ".", "INITIAL", ",", "states", ".", "SCHEDULED", "]", ":", "if", "executable_stage", ".", "state", "==", "states", ".", "INITIAL", ":", "transition", "(", "obj", "=", "executable_stage", ",", "obj_type", "=", "'Stage'", ",", "new_state", "=", "states", ".", "SCHEDULING", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "executable_tasks", "=", "executable_stage", ".", "tasks", "for", "executable_task", "in", "executable_tasks", ":", "if", "(", "executable_task", ".", "state", "==", "states", ".", "INITIAL", ")", "or", "(", "(", "executable_task", ".", "state", "==", "states", ".", "FAILED", ")", "and", "(", "self", ".", "_resubmit_failed", ")", ")", ":", "# Set state of Tasks in current Stage to SCHEDULING", "transition", "(", "obj", "=", "executable_task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "SCHEDULING", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "# task_as_dict = json.dumps(executable_task.to_dict())", "workload", ".", "append", "(", "executable_task", ")", "if", "executable_stage", "not", "in", "scheduled_stages", ":", "scheduled_stages", ".", "append", "(", "executable_stage", ")", "if", "workload", ":", "# Put the task on one of the pending_queues", "workload_as_dict", "=", "list", "(", ")", "for", "task", "in", "workload", ":", "workload_as_dict", ".", "append", "(", "task", ".", "to_dict", "(", ")", ")", "workload_as_dict", "=", "json", ".", "dumps", "(", "workload_as_dict", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "self", ".", "_pending_queue", "[", "0", "]", ",", "body", "=", "workload_as_dict", "# properties=pika.BasicProperties(", "# make message persistent", "# delivery_mode = 2)", ")", "for", "task", "in", "workload", ":", "# Set state of Tasks in current Stage to SCHEDULED", "transition", "(", "obj", "=", "task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "SCHEDULED", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "self", ".", "_logger", ".", "debug", "(", "'Task %s published to pending queue'", "%", "task", ".", "uid", ")", "if", "scheduled_stages", ":", "for", "executable_stage", "in", "scheduled_stages", ":", "transition", "(", "obj", "=", "executable_stage", ",", "obj_type", "=", "'Stage'", ",", "new_state", "=", "states", ".", "SCHEDULED", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-enq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "# Appease pika cos it thinks the connection is dead", "now", "=", "time", ".", "time", "(", ")", "if", "now", "-", "last", ">=", "self", ".", "_rmq_ping_interval", ":", "mq_connection", ".", "process_data_events", "(", ")", "last", "=", "now", "self", ".", "_logger", ".", "info", "(", "'Enqueue thread terminated'", ")", "mq_connection", ".", "close", "(", ")", "local_prof", ".", "prof", "(", "'terminating enqueue-thread'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "KeyboardInterrupt", ":", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to cancel enqueuer thread gracefully...'", ")", "mq_connection", ".", "close", "(", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Error in enqueue-thread: %s'", "%", "ex", ")", "try", ":", "mq_connection", ".", "close", "(", ")", "except", "Exception", "as", "ex", ":", "self", ".", "_logger", ".", "warning", "(", "'mq_connection not created, %s'", "%", "ex", ")", "raise" ]
**Purpose**: This is the function that is run in the enqueue thread. This function extracts Tasks from the copy of workflow that exists in the WFprocessor object and pushes them to the queues in the pending_q list. Since this thread works on the copy of the workflow, every state update to the Task, Stage and Pipeline is communicated back to the AppManager (master process) via the 'sync_with_master' function that has dedicated queues to communicate with the master. Details: Termination condition of this thread is set by the wfp process.
[ "**", "Purpose", "**", ":", "This", "is", "the", "function", "that", "is", "run", "in", "the", "enqueue", "thread", ".", "This", "function", "extracts", "Tasks", "from", "the", "copy", "of", "workflow", "that", "exists", "in", "the", "WFprocessor", "object", "and", "pushes", "them", "to", "the", "queues", "in", "the", "pending_q", "list", ".", "Since", "this", "thread", "works", "on", "the", "copy", "of", "the", "workflow", "every", "state", "update", "to", "the", "Task", "Stage", "and", "Pipeline", "is", "communicated", "back", "to", "the", "AppManager", "(", "master", "process", ")", "via", "the", "sync_with_master", "function", "that", "has", "dedicated", "queues", "to", "communicate", "with", "the", "master", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L106-L271
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor._dequeue
def _dequeue(self, local_prof): """ **Purpose**: This is the function that is run in the dequeue thread. This function extracts Tasks from the completed queus and updates the copy of workflow that exists in the WFprocessor object. Since this thread works on the copy of the workflow, every state update to the Task, Stage and Pipeline is communicated back to the AppManager (master process) via the 'sync_with_master' function that has dedicated queues to communicate with the master. Details: Termination condition of this thread is set by the wfp process. """ try: local_prof.prof('dequeue-thread started', uid=self._uid) self._logger.info('Dequeue thread started') mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() last = time.time() while not self._dequeue_thread_terminate.is_set(): try: method_frame, header_frame, body = mq_channel.basic_get( queue=self._completed_queue[0]) if body: # Get task from the message completed_task = Task() completed_task.from_dict(json.loads(body)) self._logger.info( 'Got finished task %s from queue' % (completed_task.uid)) transition(obj=completed_task, obj_type='Task', new_state=states.DEQUEUEING, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Traverse the entire workflow to find out the correct Task for pipe in self._workflow: with pipe.lock: if ((not pipe.completed) and (not pipe.state == states.SUSPENDED)): if completed_task.parent_pipeline['uid'] == pipe.uid: self._logger.debug( 'Found parent pipeline: %s' % pipe.uid) for stage in pipe.stages: if completed_task.parent_stage['uid'] == stage.uid: self._logger.debug( 'Found parent stage: %s' % (stage.uid)) transition(obj=completed_task, obj_type='Task', new_state=states.DEQUEUED, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) if not completed_task.exit_code: completed_task.state = states.DONE else: completed_task.state = states.FAILED for task in stage.tasks: if task.uid == completed_task.uid: task.state = str( completed_task.state) if (task.state == states.FAILED) and (self._resubmit_failed): task.state = states.INITIAL transition(obj=task, obj_type='Task', new_state=task.state, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) if stage._check_stage_complete(): transition(obj=stage, obj_type='Stage', new_state=states.DONE, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Check if Stage has a post-exec that needs to be # executed if stage.post_exec: try: self._logger.info('Executing post-exec for stage %s' % stage.uid) self._prof.prof('Adap: executing post-exec', uid=self._uid) stage.post_exec() self._logger.info( 'Post-exec executed for stage %s' % stage.uid) self._prof.prof( 'Adap: post-exec executed', uid=self._uid) except Exception, ex: self._logger.exception('Execution failed in post_exec of stage %s' % stage.uid) raise pipe._increment_stage() if pipe.completed: transition(obj=pipe, obj_type='Pipeline', new_state=states.DONE, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Found the task and processed it -- no more iterations needed break # Found the stage and processed it -- no more iterations neeeded break # Found the pipeline and processed it -- no more iterations neeeded break mq_channel.basic_ack( delivery_tag=method_frame.delivery_tag) # Appease pika cos it thinks the connection is dead now = time.time() if now - last >= self._rmq_ping_interval: mq_connection.process_data_events() last = now except Exception, ex: self._logger.exception( 'Unable to receive message from completed queue: %s' % ex) raise self._logger.info('Terminated dequeue thread') mq_connection.close() local_prof.prof('terminating dequeue-thread', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit gracefully...') mq_connection.close() raise KeyboardInterrupt except Exception, ex: self._logger.exception('Error in dequeue-thread: %s' % ex) try: mq_connection.close() except: self._logger.warning('mq_connection not created') raise EnTKError(ex)
python
def _dequeue(self, local_prof): """ **Purpose**: This is the function that is run in the dequeue thread. This function extracts Tasks from the completed queus and updates the copy of workflow that exists in the WFprocessor object. Since this thread works on the copy of the workflow, every state update to the Task, Stage and Pipeline is communicated back to the AppManager (master process) via the 'sync_with_master' function that has dedicated queues to communicate with the master. Details: Termination condition of this thread is set by the wfp process. """ try: local_prof.prof('dequeue-thread started', uid=self._uid) self._logger.info('Dequeue thread started') mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=self._mq_hostname, port=self._port)) mq_channel = mq_connection.channel() last = time.time() while not self._dequeue_thread_terminate.is_set(): try: method_frame, header_frame, body = mq_channel.basic_get( queue=self._completed_queue[0]) if body: # Get task from the message completed_task = Task() completed_task.from_dict(json.loads(body)) self._logger.info( 'Got finished task %s from queue' % (completed_task.uid)) transition(obj=completed_task, obj_type='Task', new_state=states.DEQUEUEING, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Traverse the entire workflow to find out the correct Task for pipe in self._workflow: with pipe.lock: if ((not pipe.completed) and (not pipe.state == states.SUSPENDED)): if completed_task.parent_pipeline['uid'] == pipe.uid: self._logger.debug( 'Found parent pipeline: %s' % pipe.uid) for stage in pipe.stages: if completed_task.parent_stage['uid'] == stage.uid: self._logger.debug( 'Found parent stage: %s' % (stage.uid)) transition(obj=completed_task, obj_type='Task', new_state=states.DEQUEUED, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) if not completed_task.exit_code: completed_task.state = states.DONE else: completed_task.state = states.FAILED for task in stage.tasks: if task.uid == completed_task.uid: task.state = str( completed_task.state) if (task.state == states.FAILED) and (self._resubmit_failed): task.state = states.INITIAL transition(obj=task, obj_type='Task', new_state=task.state, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) if stage._check_stage_complete(): transition(obj=stage, obj_type='Stage', new_state=states.DONE, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Check if Stage has a post-exec that needs to be # executed if stage.post_exec: try: self._logger.info('Executing post-exec for stage %s' % stage.uid) self._prof.prof('Adap: executing post-exec', uid=self._uid) stage.post_exec() self._logger.info( 'Post-exec executed for stage %s' % stage.uid) self._prof.prof( 'Adap: post-exec executed', uid=self._uid) except Exception, ex: self._logger.exception('Execution failed in post_exec of stage %s' % stage.uid) raise pipe._increment_stage() if pipe.completed: transition(obj=pipe, obj_type='Pipeline', new_state=states.DONE, channel=mq_channel, queue='%s-deq-to-sync' % self._sid, profiler=local_prof, logger=self._logger) # Found the task and processed it -- no more iterations needed break # Found the stage and processed it -- no more iterations neeeded break # Found the pipeline and processed it -- no more iterations neeeded break mq_channel.basic_ack( delivery_tag=method_frame.delivery_tag) # Appease pika cos it thinks the connection is dead now = time.time() if now - last >= self._rmq_ping_interval: mq_connection.process_data_events() last = now except Exception, ex: self._logger.exception( 'Unable to receive message from completed queue: %s' % ex) raise self._logger.info('Terminated dequeue thread') mq_connection.close() local_prof.prof('terminating dequeue-thread', uid=self._uid) except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit gracefully...') mq_connection.close() raise KeyboardInterrupt except Exception, ex: self._logger.exception('Error in dequeue-thread: %s' % ex) try: mq_connection.close() except: self._logger.warning('mq_connection not created') raise EnTKError(ex)
[ "def", "_dequeue", "(", "self", ",", "local_prof", ")", ":", "try", ":", "local_prof", ".", "prof", "(", "'dequeue-thread started'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'Dequeue thread started'", ")", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "self", ".", "_mq_hostname", ",", "port", "=", "self", ".", "_port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "last", "=", "time", ".", "time", "(", ")", "while", "not", "self", ".", "_dequeue_thread_terminate", ".", "is_set", "(", ")", ":", "try", ":", "method_frame", ",", "header_frame", ",", "body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "self", ".", "_completed_queue", "[", "0", "]", ")", "if", "body", ":", "# Get task from the message", "completed_task", "=", "Task", "(", ")", "completed_task", ".", "from_dict", "(", "json", ".", "loads", "(", "body", ")", ")", "self", ".", "_logger", ".", "info", "(", "'Got finished task %s from queue'", "%", "(", "completed_task", ".", "uid", ")", ")", "transition", "(", "obj", "=", "completed_task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "DEQUEUEING", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "# Traverse the entire workflow to find out the correct Task", "for", "pipe", "in", "self", ".", "_workflow", ":", "with", "pipe", ".", "lock", ":", "if", "(", "(", "not", "pipe", ".", "completed", ")", "and", "(", "not", "pipe", ".", "state", "==", "states", ".", "SUSPENDED", ")", ")", ":", "if", "completed_task", ".", "parent_pipeline", "[", "'uid'", "]", "==", "pipe", ".", "uid", ":", "self", ".", "_logger", ".", "debug", "(", "'Found parent pipeline: %s'", "%", "pipe", ".", "uid", ")", "for", "stage", "in", "pipe", ".", "stages", ":", "if", "completed_task", ".", "parent_stage", "[", "'uid'", "]", "==", "stage", ".", "uid", ":", "self", ".", "_logger", ".", "debug", "(", "'Found parent stage: %s'", "%", "(", "stage", ".", "uid", ")", ")", "transition", "(", "obj", "=", "completed_task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "DEQUEUED", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "if", "not", "completed_task", ".", "exit_code", ":", "completed_task", ".", "state", "=", "states", ".", "DONE", "else", ":", "completed_task", ".", "state", "=", "states", ".", "FAILED", "for", "task", "in", "stage", ".", "tasks", ":", "if", "task", ".", "uid", "==", "completed_task", ".", "uid", ":", "task", ".", "state", "=", "str", "(", "completed_task", ".", "state", ")", "if", "(", "task", ".", "state", "==", "states", ".", "FAILED", ")", "and", "(", "self", ".", "_resubmit_failed", ")", ":", "task", ".", "state", "=", "states", ".", "INITIAL", "transition", "(", "obj", "=", "task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "task", ".", "state", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "if", "stage", ".", "_check_stage_complete", "(", ")", ":", "transition", "(", "obj", "=", "stage", ",", "obj_type", "=", "'Stage'", ",", "new_state", "=", "states", ".", "DONE", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "# Check if Stage has a post-exec that needs to be", "# executed", "if", "stage", ".", "post_exec", ":", "try", ":", "self", ".", "_logger", ".", "info", "(", "'Executing post-exec for stage %s'", "%", "stage", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'Adap: executing post-exec'", ",", "uid", "=", "self", ".", "_uid", ")", "stage", ".", "post_exec", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Post-exec executed for stage %s'", "%", "stage", ".", "uid", ")", "self", ".", "_prof", ".", "prof", "(", "'Adap: post-exec executed'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Execution failed in post_exec of stage %s'", "%", "stage", ".", "uid", ")", "raise", "pipe", ".", "_increment_stage", "(", ")", "if", "pipe", ".", "completed", ":", "transition", "(", "obj", "=", "pipe", ",", "obj_type", "=", "'Pipeline'", ",", "new_state", "=", "states", ".", "DONE", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-deq-to-sync'", "%", "self", ".", "_sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "self", ".", "_logger", ")", "# Found the task and processed it -- no more iterations needed", "break", "# Found the stage and processed it -- no more iterations neeeded", "break", "# Found the pipeline and processed it -- no more iterations neeeded", "break", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "# Appease pika cos it thinks the connection is dead", "now", "=", "time", ".", "time", "(", ")", "if", "now", "-", "last", ">=", "self", ".", "_rmq_ping_interval", ":", "mq_connection", ".", "process_data_events", "(", ")", "last", "=", "now", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Unable to receive message from completed queue: %s'", "%", "ex", ")", "raise", "self", ".", "_logger", ".", "info", "(", "'Terminated dequeue thread'", ")", "mq_connection", ".", "close", "(", ")", "local_prof", ".", "prof", "(", "'terminating dequeue-thread'", ",", "uid", "=", "self", ".", "_uid", ")", "except", "KeyboardInterrupt", ":", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to exit gracefully...'", ")", "mq_connection", ".", "close", "(", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Error in dequeue-thread: %s'", "%", "ex", ")", "try", ":", "mq_connection", ".", "close", "(", ")", "except", ":", "self", ".", "_logger", ".", "warning", "(", "'mq_connection not created'", ")", "raise", "EnTKError", "(", "ex", ")" ]
**Purpose**: This is the function that is run in the dequeue thread. This function extracts Tasks from the completed queus and updates the copy of workflow that exists in the WFprocessor object. Since this thread works on the copy of the workflow, every state update to the Task, Stage and Pipeline is communicated back to the AppManager (master process) via the 'sync_with_master' function that has dedicated queues to communicate with the master. Details: Termination condition of this thread is set by the wfp process.
[ "**", "Purpose", "**", ":", "This", "is", "the", "function", "that", "is", "run", "in", "the", "dequeue", "thread", ".", "This", "function", "extracts", "Tasks", "from", "the", "completed", "queus", "and", "updates", "the", "copy", "of", "workflow", "that", "exists", "in", "the", "WFprocessor", "object", ".", "Since", "this", "thread", "works", "on", "the", "copy", "of", "the", "workflow", "every", "state", "update", "to", "the", "Task", "Stage", "and", "Pipeline", "is", "communicated", "back", "to", "the", "AppManager", "(", "master", "process", ")", "via", "the", "sync_with_master", "function", "that", "has", "dedicated", "queues", "to", "communicate", "with", "the", "master", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L273-L456
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor._wfp
def _wfp(self): """ **Purpose**: This is the function executed in the wfp process. The function is used to simply create and spawn two threads: enqueue, dequeue. The enqueue thread pushes ready tasks to the queues in the pending_q slow list whereas the dequeue thread pulls completed tasks from the queues in the completed_q. This function is also responsible for the termination of these threads and hence blocking. """ try: local_prof = ru.Profiler( name='radical.entk.%s' % self._uid + '-proc', path=self._path) local_prof.prof('wfp process started', uid=self._uid) self._logger.info('WFprocessor started') # Process should run till terminate condtion is encountered while (not self._wfp_terminate.is_set()): try: # Start dequeue thread if (not self._dequeue_thread) or (not self._dequeue_thread.is_alive()): local_prof.prof( 'creating dequeue-thread', uid=self._uid) self._dequeue_thread = threading.Thread( target=self._dequeue, args=(local_prof,), name='dequeue-thread') self._logger.info('Starting dequeue-thread') local_prof.prof( 'starting dequeue-thread', uid=self._uid) self._dequeue_thread.start() # Start enqueue thread if (not self._enqueue_thread) or (not self._enqueue_thread.is_alive()): local_prof.prof( 'creating enqueue-thread', uid=self._uid) self._enqueue_thread = threading.Thread( target=self._enqueue, args=(local_prof,), name='enqueue-thread') self._logger.info('Starting enqueue-thread') local_prof.prof( 'starting enqueue-thread', uid=self._uid) self._enqueue_thread.start() except Exception, ex: self._logger.exception('WFProcessor interrupted') raise local_prof.prof('start termination', uid=self._uid) self._logger.info('Terminating enqueue-thread') self._enqueue_thread_terminate.set() self._enqueue_thread.join() self._logger.info('Terminating dequeue-thread') self._dequeue_thread_terminate.set() self._dequeue_thread.join() local_prof.prof('termination done', uid=self._uid) local_prof.prof('terminating wfp process', uid=self._uid) local_prof.close() except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel wfprocessor process gracefully...') if self._enqueue_thread: if not self._enqueue_thread_terminate.is_set(): self._logger.info('Terminating enqueue-thread') self._enqueue_thread_terminate.set() self._enqueue_thread.join() if self._dequeue_thread: if not self._dequeue_thread_terminate.is_set(): self._logger.info('Terminating dequeue-thread') self._dequeue_thread_terminate.set() self._dequeue_thread.join() self._logger.info('WFprocessor process terminated') raise KeyboardInterrupt except Exception, ex: self._logger.exception( 'Error in wfp process: %s. \n Closing enqueue, dequeue threads' % ex) if self._enqueue_thread: if not self._enqueue_thread_terminate.is_set(): self._logger.info('Terminating enqueue-thread') self._enqueue_thread_terminate.set() self._enqueue_thread.join() if self._dequeue_thread: if not self._dequeue_thread_terminate.is_set(): self._logger.info('Terminating dequeue-thread') self._dequeue_thread_terminate.set() self._dequeue_thread.join() self._logger.info('WFprocessor process terminated') self._logger.exception('%s failed with %s'%(self._uid, ex)) raise EnTKError(ex)
python
def _wfp(self): """ **Purpose**: This is the function executed in the wfp process. The function is used to simply create and spawn two threads: enqueue, dequeue. The enqueue thread pushes ready tasks to the queues in the pending_q slow list whereas the dequeue thread pulls completed tasks from the queues in the completed_q. This function is also responsible for the termination of these threads and hence blocking. """ try: local_prof = ru.Profiler( name='radical.entk.%s' % self._uid + '-proc', path=self._path) local_prof.prof('wfp process started', uid=self._uid) self._logger.info('WFprocessor started') # Process should run till terminate condtion is encountered while (not self._wfp_terminate.is_set()): try: # Start dequeue thread if (not self._dequeue_thread) or (not self._dequeue_thread.is_alive()): local_prof.prof( 'creating dequeue-thread', uid=self._uid) self._dequeue_thread = threading.Thread( target=self._dequeue, args=(local_prof,), name='dequeue-thread') self._logger.info('Starting dequeue-thread') local_prof.prof( 'starting dequeue-thread', uid=self._uid) self._dequeue_thread.start() # Start enqueue thread if (not self._enqueue_thread) or (not self._enqueue_thread.is_alive()): local_prof.prof( 'creating enqueue-thread', uid=self._uid) self._enqueue_thread = threading.Thread( target=self._enqueue, args=(local_prof,), name='enqueue-thread') self._logger.info('Starting enqueue-thread') local_prof.prof( 'starting enqueue-thread', uid=self._uid) self._enqueue_thread.start() except Exception, ex: self._logger.exception('WFProcessor interrupted') raise local_prof.prof('start termination', uid=self._uid) self._logger.info('Terminating enqueue-thread') self._enqueue_thread_terminate.set() self._enqueue_thread.join() self._logger.info('Terminating dequeue-thread') self._dequeue_thread_terminate.set() self._dequeue_thread.join() local_prof.prof('termination done', uid=self._uid) local_prof.prof('terminating wfp process', uid=self._uid) local_prof.close() except KeyboardInterrupt: self._logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel wfprocessor process gracefully...') if self._enqueue_thread: if not self._enqueue_thread_terminate.is_set(): self._logger.info('Terminating enqueue-thread') self._enqueue_thread_terminate.set() self._enqueue_thread.join() if self._dequeue_thread: if not self._dequeue_thread_terminate.is_set(): self._logger.info('Terminating dequeue-thread') self._dequeue_thread_terminate.set() self._dequeue_thread.join() self._logger.info('WFprocessor process terminated') raise KeyboardInterrupt except Exception, ex: self._logger.exception( 'Error in wfp process: %s. \n Closing enqueue, dequeue threads' % ex) if self._enqueue_thread: if not self._enqueue_thread_terminate.is_set(): self._logger.info('Terminating enqueue-thread') self._enqueue_thread_terminate.set() self._enqueue_thread.join() if self._dequeue_thread: if not self._dequeue_thread_terminate.is_set(): self._logger.info('Terminating dequeue-thread') self._dequeue_thread_terminate.set() self._dequeue_thread.join() self._logger.info('WFprocessor process terminated') self._logger.exception('%s failed with %s'%(self._uid, ex)) raise EnTKError(ex)
[ "def", "_wfp", "(", "self", ")", ":", "try", ":", "local_prof", "=", "ru", ".", "Profiler", "(", "name", "=", "'radical.entk.%s'", "%", "self", ".", "_uid", "+", "'-proc'", ",", "path", "=", "self", ".", "_path", ")", "local_prof", ".", "prof", "(", "'wfp process started'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'WFprocessor started'", ")", "# Process should run till terminate condtion is encountered", "while", "(", "not", "self", ".", "_wfp_terminate", ".", "is_set", "(", ")", ")", ":", "try", ":", "# Start dequeue thread", "if", "(", "not", "self", ".", "_dequeue_thread", ")", "or", "(", "not", "self", ".", "_dequeue_thread", ".", "is_alive", "(", ")", ")", ":", "local_prof", ".", "prof", "(", "'creating dequeue-thread'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_dequeue_thread", "=", "threading", ".", "Thread", "(", "target", "=", "self", ".", "_dequeue", ",", "args", "=", "(", "local_prof", ",", ")", ",", "name", "=", "'dequeue-thread'", ")", "self", ".", "_logger", ".", "info", "(", "'Starting dequeue-thread'", ")", "local_prof", ".", "prof", "(", "'starting dequeue-thread'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_dequeue_thread", ".", "start", "(", ")", "# Start enqueue thread", "if", "(", "not", "self", ".", "_enqueue_thread", ")", "or", "(", "not", "self", ".", "_enqueue_thread", ".", "is_alive", "(", ")", ")", ":", "local_prof", ".", "prof", "(", "'creating enqueue-thread'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_enqueue_thread", "=", "threading", ".", "Thread", "(", "target", "=", "self", ".", "_enqueue", ",", "args", "=", "(", "local_prof", ",", ")", ",", "name", "=", "'enqueue-thread'", ")", "self", ".", "_logger", ".", "info", "(", "'Starting enqueue-thread'", ")", "local_prof", ".", "prof", "(", "'starting enqueue-thread'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_enqueue_thread", ".", "start", "(", ")", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'WFProcessor interrupted'", ")", "raise", "local_prof", ".", "prof", "(", "'start termination'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_logger", ".", "info", "(", "'Terminating enqueue-thread'", ")", "self", ".", "_enqueue_thread_terminate", ".", "set", "(", ")", "self", ".", "_enqueue_thread", ".", "join", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Terminating dequeue-thread'", ")", "self", ".", "_dequeue_thread_terminate", ".", "set", "(", ")", "self", ".", "_dequeue_thread", ".", "join", "(", ")", "local_prof", ".", "prof", "(", "'termination done'", ",", "uid", "=", "self", ".", "_uid", ")", "local_prof", ".", "prof", "(", "'terminating wfp process'", ",", "uid", "=", "self", ".", "_uid", ")", "local_prof", ".", "close", "(", ")", "except", "KeyboardInterrupt", ":", "self", ".", "_logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to cancel wfprocessor process gracefully...'", ")", "if", "self", ".", "_enqueue_thread", ":", "if", "not", "self", ".", "_enqueue_thread_terminate", ".", "is_set", "(", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating enqueue-thread'", ")", "self", ".", "_enqueue_thread_terminate", ".", "set", "(", ")", "self", ".", "_enqueue_thread", ".", "join", "(", ")", "if", "self", ".", "_dequeue_thread", ":", "if", "not", "self", ".", "_dequeue_thread_terminate", ".", "is_set", "(", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating dequeue-thread'", ")", "self", ".", "_dequeue_thread_terminate", ".", "set", "(", ")", "self", ".", "_dequeue_thread", ".", "join", "(", ")", "self", ".", "_logger", ".", "info", "(", "'WFprocessor process terminated'", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Error in wfp process: %s. \\n Closing enqueue, dequeue threads'", "%", "ex", ")", "if", "self", ".", "_enqueue_thread", ":", "if", "not", "self", ".", "_enqueue_thread_terminate", ".", "is_set", "(", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating enqueue-thread'", ")", "self", ".", "_enqueue_thread_terminate", ".", "set", "(", ")", "self", ".", "_enqueue_thread", ".", "join", "(", ")", "if", "self", ".", "_dequeue_thread", ":", "if", "not", "self", ".", "_dequeue_thread_terminate", ".", "is_set", "(", ")", ":", "self", ".", "_logger", ".", "info", "(", "'Terminating dequeue-thread'", ")", "self", ".", "_dequeue_thread_terminate", ".", "set", "(", ")", "self", ".", "_dequeue_thread", ".", "join", "(", ")", "self", ".", "_logger", ".", "info", "(", "'WFprocessor process terminated'", ")", "self", ".", "_logger", ".", "exception", "(", "'%s failed with %s'", "%", "(", "self", ".", "_uid", ",", "ex", ")", ")", "raise", "EnTKError", "(", "ex", ")" ]
**Purpose**: This is the function executed in the wfp process. The function is used to simply create and spawn two threads: enqueue, dequeue. The enqueue thread pushes ready tasks to the queues in the pending_q slow list whereas the dequeue thread pulls completed tasks from the queues in the completed_q. This function is also responsible for the termination of these threads and hence blocking.
[ "**", "Purpose", "**", ":", "This", "is", "the", "function", "executed", "in", "the", "wfp", "process", ".", "The", "function", "is", "used", "to", "simply", "create", "and", "spawn", "two", "threads", ":", "enqueue", "dequeue", ".", "The", "enqueue", "thread", "pushes", "ready", "tasks", "to", "the", "queues", "in", "the", "pending_q", "slow", "list", "whereas", "the", "dequeue", "thread", "pulls", "completed", "tasks", "from", "the", "queues", "in", "the", "completed_q", ".", "This", "function", "is", "also", "responsible", "for", "the", "termination", "of", "these", "threads", "and", "hence", "blocking", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L458-L569
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor.start_processor
def start_processor(self): """ **Purpose**: Method to start the wfp process. The wfp function is not to be accessed directly. The function is started in a separate process using this method. """ if not self._wfp_process: try: self._prof.prof('creating wfp process', uid=self._uid) self._wfp_process = Process( target=self._wfp, name='wfprocessor') self._enqueue_thread = None self._dequeue_thread = None self._enqueue_thread_terminate = threading.Event() self._dequeue_thread_terminate = threading.Event() self._wfp_terminate = Event() self._logger.info('Starting WFprocessor process') self._prof.prof('starting wfp process', uid=self._uid) self._wfp_process.start() return True except Exception, ex: self._logger.exception('WFprocessor not started') self.terminate_processor() raise else: self._logger.warn( 'Wfp process already running, attempted to restart!')
python
def start_processor(self): """ **Purpose**: Method to start the wfp process. The wfp function is not to be accessed directly. The function is started in a separate process using this method. """ if not self._wfp_process: try: self._prof.prof('creating wfp process', uid=self._uid) self._wfp_process = Process( target=self._wfp, name='wfprocessor') self._enqueue_thread = None self._dequeue_thread = None self._enqueue_thread_terminate = threading.Event() self._dequeue_thread_terminate = threading.Event() self._wfp_terminate = Event() self._logger.info('Starting WFprocessor process') self._prof.prof('starting wfp process', uid=self._uid) self._wfp_process.start() return True except Exception, ex: self._logger.exception('WFprocessor not started') self.terminate_processor() raise else: self._logger.warn( 'Wfp process already running, attempted to restart!')
[ "def", "start_processor", "(", "self", ")", ":", "if", "not", "self", ".", "_wfp_process", ":", "try", ":", "self", ".", "_prof", ".", "prof", "(", "'creating wfp process'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_wfp_process", "=", "Process", "(", "target", "=", "self", ".", "_wfp", ",", "name", "=", "'wfprocessor'", ")", "self", ".", "_enqueue_thread", "=", "None", "self", ".", "_dequeue_thread", "=", "None", "self", ".", "_enqueue_thread_terminate", "=", "threading", ".", "Event", "(", ")", "self", ".", "_dequeue_thread_terminate", "=", "threading", ".", "Event", "(", ")", "self", ".", "_wfp_terminate", "=", "Event", "(", ")", "self", ".", "_logger", ".", "info", "(", "'Starting WFprocessor process'", ")", "self", ".", "_prof", ".", "prof", "(", "'starting wfp process'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_wfp_process", ".", "start", "(", ")", "return", "True", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'WFprocessor not started'", ")", "self", ".", "terminate_processor", "(", ")", "raise", "else", ":", "self", ".", "_logger", ".", "warn", "(", "'Wfp process already running, attempted to restart!'", ")" ]
**Purpose**: Method to start the wfp process. The wfp function is not to be accessed directly. The function is started in a separate process using this method.
[ "**", "Purpose", "**", ":", "Method", "to", "start", "the", "wfp", "process", ".", "The", "wfp", "function", "is", "not", "to", "be", "accessed", "directly", ".", "The", "function", "is", "started", "in", "a", "separate", "process", "using", "this", "method", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L575-L610
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor.terminate_processor
def terminate_processor(self): """ **Purpose**: Method to terminate the wfp process. This method is blocking as it waits for the wfp process to terminate (aka join). """ try: if self.check_processor(): self._logger.debug( 'Attempting to end WFprocessor... event: %s' % self._wfp_terminate.is_set()) self._wfp_terminate.set() self._wfp_process.join() self._wfp_process = None self._logger.debug('WFprocessor process terminated') else: self._logger.debug('WFprocessor process already terminated') self._prof.prof('wfp process terminated', uid=self._uid) self._prof.close() except Exception, ex: self._logger.exception('Could not terminate wfprocessor process') raise
python
def terminate_processor(self): """ **Purpose**: Method to terminate the wfp process. This method is blocking as it waits for the wfp process to terminate (aka join). """ try: if self.check_processor(): self._logger.debug( 'Attempting to end WFprocessor... event: %s' % self._wfp_terminate.is_set()) self._wfp_terminate.set() self._wfp_process.join() self._wfp_process = None self._logger.debug('WFprocessor process terminated') else: self._logger.debug('WFprocessor process already terminated') self._prof.prof('wfp process terminated', uid=self._uid) self._prof.close() except Exception, ex: self._logger.exception('Could not terminate wfprocessor process') raise
[ "def", "terminate_processor", "(", "self", ")", ":", "try", ":", "if", "self", ".", "check_processor", "(", ")", ":", "self", ".", "_logger", ".", "debug", "(", "'Attempting to end WFprocessor... event: %s'", "%", "self", ".", "_wfp_terminate", ".", "is_set", "(", ")", ")", "self", ".", "_wfp_terminate", ".", "set", "(", ")", "self", ".", "_wfp_process", ".", "join", "(", ")", "self", ".", "_wfp_process", "=", "None", "self", ".", "_logger", ".", "debug", "(", "'WFprocessor process terminated'", ")", "else", ":", "self", ".", "_logger", ".", "debug", "(", "'WFprocessor process already terminated'", ")", "self", ".", "_prof", ".", "prof", "(", "'wfp process terminated'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_prof", ".", "close", "(", ")", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Could not terminate wfprocessor process'", ")", "raise" ]
**Purpose**: Method to terminate the wfp process. This method is blocking as it waits for the wfp process to terminate (aka join).
[ "**", "Purpose", "**", ":", "Method", "to", "terminate", "the", "wfp", "process", ".", "This", "method", "is", "blocking", "as", "it", "waits", "for", "the", "wfp", "process", "to", "terminate", "(", "aka", "join", ")", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L612-L636
radical-cybertools/radical.entk
src/radical/entk/appman/wfprocessor.py
WFprocessor.workflow_incomplete
def workflow_incomplete(self): """ **Purpose**: Method to check if the workflow execution is incomplete. """ try: for pipe in self._workflow: with pipe.lock: if pipe.completed: pass else: return True return False except Exception, ex: self._logger.exception( 'Could not check if workflow is incomplete, error:%s' % ex) raise
python
def workflow_incomplete(self): """ **Purpose**: Method to check if the workflow execution is incomplete. """ try: for pipe in self._workflow: with pipe.lock: if pipe.completed: pass else: return True return False except Exception, ex: self._logger.exception( 'Could not check if workflow is incomplete, error:%s' % ex) raise
[ "def", "workflow_incomplete", "(", "self", ")", ":", "try", ":", "for", "pipe", "in", "self", ".", "_workflow", ":", "with", "pipe", ".", "lock", ":", "if", "pipe", ".", "completed", ":", "pass", "else", ":", "return", "True", "return", "False", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Could not check if workflow is incomplete, error:%s'", "%", "ex", ")", "raise" ]
**Purpose**: Method to check if the workflow execution is incomplete.
[ "**", "Purpose", "**", ":", "Method", "to", "check", "if", "the", "workflow", "execution", "is", "incomplete", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/appman/wfprocessor.py#L638-L655
google/dotty
efilter/protocols/repeated.py
meld
def meld(*values): """Return the repeated value, or the first value if there's only one. This is a convenience function, equivalent to calling getvalue(repeated(x)) to get x. This function skips over instances of None in values (None is not allowed in repeated variables). Examples: meld("foo", "bar") # => ListRepetition("foo", "bar") meld("foo", "foo") # => ListRepetition("foo", "foo") meld("foo", None) # => "foo" meld(None) # => None """ values = [x for x in values if x is not None] if not values: return None result = repeated(*values) if isrepeating(result): return result return getvalue(result)
python
def meld(*values): """Return the repeated value, or the first value if there's only one. This is a convenience function, equivalent to calling getvalue(repeated(x)) to get x. This function skips over instances of None in values (None is not allowed in repeated variables). Examples: meld("foo", "bar") # => ListRepetition("foo", "bar") meld("foo", "foo") # => ListRepetition("foo", "foo") meld("foo", None) # => "foo" meld(None) # => None """ values = [x for x in values if x is not None] if not values: return None result = repeated(*values) if isrepeating(result): return result return getvalue(result)
[ "def", "meld", "(", "*", "values", ")", ":", "values", "=", "[", "x", "for", "x", "in", "values", "if", "x", "is", "not", "None", "]", "if", "not", "values", ":", "return", "None", "result", "=", "repeated", "(", "*", "values", ")", "if", "isrepeating", "(", "result", ")", ":", "return", "result", "return", "getvalue", "(", "result", ")" ]
Return the repeated value, or the first value if there's only one. This is a convenience function, equivalent to calling getvalue(repeated(x)) to get x. This function skips over instances of None in values (None is not allowed in repeated variables). Examples: meld("foo", "bar") # => ListRepetition("foo", "bar") meld("foo", "foo") # => ListRepetition("foo", "foo") meld("foo", None) # => "foo" meld(None) # => None
[ "Return", "the", "repeated", "value", "or", "the", "first", "value", "if", "there", "s", "only", "one", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/protocols/repeated.py#L55-L78
google/dotty
efilter/protocols/repeated.py
getvalue
def getvalue(x): """Return the single value of x or raise TypError if more than one value.""" if isrepeating(x): raise TypeError( "Ambiguous call to getvalue for %r which has more than one value." % x) for value in getvalues(x): return value
python
def getvalue(x): """Return the single value of x or raise TypError if more than one value.""" if isrepeating(x): raise TypeError( "Ambiguous call to getvalue for %r which has more than one value." % x) for value in getvalues(x): return value
[ "def", "getvalue", "(", "x", ")", ":", "if", "isrepeating", "(", "x", ")", ":", "raise", "TypeError", "(", "\"Ambiguous call to getvalue for %r which has more than one value.\"", "%", "x", ")", "for", "value", "in", "getvalues", "(", "x", ")", ":", "return", "value" ]
Return the single value of x or raise TypError if more than one value.
[ "Return", "the", "single", "value", "of", "x", "or", "raise", "TypError", "if", "more", "than", "one", "value", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/protocols/repeated.py#L120-L128
radical-cybertools/radical.entk
src/radical/entk/task/task.py
Task.luid
def luid(self): """ Unique ID of the current task (fully qualified). example: >>> task.luid pipe.0001.stage.0004.task.0234 :getter: Returns the fully qualified uid of the current task :type: String """ p_elem = self.parent_pipeline.get('name') if not p_elem: p_elem = self.parent_pipeline['uid'] s_elem = self.parent_stage.get('name') if not s_elem: s_elem = self.parent_stage['uid'] t_elem = self.name if not t_elem: t_elem = self.uid return '%s.%s.%s' % (p_elem, s_elem, t_elem)
python
def luid(self): """ Unique ID of the current task (fully qualified). example: >>> task.luid pipe.0001.stage.0004.task.0234 :getter: Returns the fully qualified uid of the current task :type: String """ p_elem = self.parent_pipeline.get('name') if not p_elem: p_elem = self.parent_pipeline['uid'] s_elem = self.parent_stage.get('name') if not s_elem: s_elem = self.parent_stage['uid'] t_elem = self.name if not t_elem: t_elem = self.uid return '%s.%s.%s' % (p_elem, s_elem, t_elem)
[ "def", "luid", "(", "self", ")", ":", "p_elem", "=", "self", ".", "parent_pipeline", ".", "get", "(", "'name'", ")", "if", "not", "p_elem", ":", "p_elem", "=", "self", ".", "parent_pipeline", "[", "'uid'", "]", "s_elem", "=", "self", ".", "parent_stage", ".", "get", "(", "'name'", ")", "if", "not", "s_elem", ":", "s_elem", "=", "self", ".", "parent_stage", "[", "'uid'", "]", "t_elem", "=", "self", ".", "name", "if", "not", "t_elem", ":", "t_elem", "=", "self", ".", "uid", "return", "'%s.%s.%s'", "%", "(", "p_elem", ",", "s_elem", ",", "t_elem", ")" ]
Unique ID of the current task (fully qualified). example: >>> task.luid pipe.0001.stage.0004.task.0234 :getter: Returns the fully qualified uid of the current task :type: String
[ "Unique", "ID", "of", "the", "current", "task", "(", "fully", "qualified", ")", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/task/task.py#L88-L111
radical-cybertools/radical.entk
src/radical/entk/task/task.py
Task.to_dict
def to_dict(self): """ Convert current Task into a dictionary :return: python dictionary """ task_desc_as_dict = { 'uid': self._uid, 'name': self._name, 'state': self._state, 'state_history': self._state_history, 'pre_exec': self._pre_exec, 'executable': self._executable, 'arguments': self._arguments, 'post_exec': self._post_exec, 'cpu_reqs': self._cpu_reqs, 'gpu_reqs': self._gpu_reqs, 'lfs_per_process': self._lfs_per_process, 'upload_input_data': self._upload_input_data, 'copy_input_data': self._copy_input_data, 'link_input_data': self._link_input_data, 'move_input_data': self._move_input_data, 'copy_output_data': self._copy_output_data, 'move_output_data': self._move_output_data, 'download_output_data': self._download_output_data, 'stdout': self._stdout, 'stderr': self._stderr, 'exit_code': self._exit_code, 'path': self._path, 'tag': self._tag, 'parent_stage': self._p_stage, 'parent_pipeline': self._p_pipeline, } return task_desc_as_dict
python
def to_dict(self): """ Convert current Task into a dictionary :return: python dictionary """ task_desc_as_dict = { 'uid': self._uid, 'name': self._name, 'state': self._state, 'state_history': self._state_history, 'pre_exec': self._pre_exec, 'executable': self._executable, 'arguments': self._arguments, 'post_exec': self._post_exec, 'cpu_reqs': self._cpu_reqs, 'gpu_reqs': self._gpu_reqs, 'lfs_per_process': self._lfs_per_process, 'upload_input_data': self._upload_input_data, 'copy_input_data': self._copy_input_data, 'link_input_data': self._link_input_data, 'move_input_data': self._move_input_data, 'copy_output_data': self._copy_output_data, 'move_output_data': self._move_output_data, 'download_output_data': self._download_output_data, 'stdout': self._stdout, 'stderr': self._stderr, 'exit_code': self._exit_code, 'path': self._path, 'tag': self._tag, 'parent_stage': self._p_stage, 'parent_pipeline': self._p_pipeline, } return task_desc_as_dict
[ "def", "to_dict", "(", "self", ")", ":", "task_desc_as_dict", "=", "{", "'uid'", ":", "self", ".", "_uid", ",", "'name'", ":", "self", ".", "_name", ",", "'state'", ":", "self", ".", "_state", ",", "'state_history'", ":", "self", ".", "_state_history", ",", "'pre_exec'", ":", "self", ".", "_pre_exec", ",", "'executable'", ":", "self", ".", "_executable", ",", "'arguments'", ":", "self", ".", "_arguments", ",", "'post_exec'", ":", "self", ".", "_post_exec", ",", "'cpu_reqs'", ":", "self", ".", "_cpu_reqs", ",", "'gpu_reqs'", ":", "self", ".", "_gpu_reqs", ",", "'lfs_per_process'", ":", "self", ".", "_lfs_per_process", ",", "'upload_input_data'", ":", "self", ".", "_upload_input_data", ",", "'copy_input_data'", ":", "self", ".", "_copy_input_data", ",", "'link_input_data'", ":", "self", ".", "_link_input_data", ",", "'move_input_data'", ":", "self", ".", "_move_input_data", ",", "'copy_output_data'", ":", "self", ".", "_copy_output_data", ",", "'move_output_data'", ":", "self", ".", "_move_output_data", ",", "'download_output_data'", ":", "self", ".", "_download_output_data", ",", "'stdout'", ":", "self", ".", "_stdout", ",", "'stderr'", ":", "self", ".", "_stderr", ",", "'exit_code'", ":", "self", ".", "_exit_code", ",", "'path'", ":", "self", ".", "_path", ",", "'tag'", ":", "self", ".", "_tag", ",", "'parent_stage'", ":", "self", ".", "_p_stage", ",", "'parent_pipeline'", ":", "self", ".", "_p_pipeline", ",", "}", "return", "task_desc_as_dict" ]
Convert current Task into a dictionary :return: python dictionary
[ "Convert", "current", "Task", "into", "a", "dictionary" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/task/task.py#L724-L764
radical-cybertools/radical.entk
src/radical/entk/task/task.py
Task.from_dict
def from_dict(self, d): """ Create a Task from a dictionary. The change is in inplace. :argument: python dictionary :return: None """ if 'uid' in d: if d['uid']: self._uid = d['uid'] if 'name' in d: if d['name']: self._name = d['name'] if 'state' in d: if isinstance(d['state'], str) or isinstance(d['state'], unicode): self._state = d['state'] else: raise TypeError(entity='state', expected_type=str, actual_type=type(d['state'])) else: self._state = states.INITIAL if 'state_history' in d: if isinstance(d['state_history'], list): self._state_history = d['state_history'] else: raise TypeError(entity='state_history', expected_type=list, actual_type=type( d['state_history'])) if 'pre_exec' in d: if isinstance(d['pre_exec'], list): self._pre_exec = d['pre_exec'] else: raise TypeError(expected_type=list, actual_type=type(d['pre_exec'])) if 'executable' in d: if isinstance(d['executable'], str) or isinstance(d['executable'], unicode): self._executable = d['executable'] else: raise TypeError(expected_type=str, actual_type=type(d['executable'])) if 'arguments' in d: if isinstance(d['arguments'], list): self._arguments = d['arguments'] else: raise TypeError(expected_type=list, actual_type=type(d['arguments'])) if 'post_exec' in d: if isinstance(d['post_exec'], list): self._post_exec = d['post_exec'] else: raise TypeError(expected_type=list, actual_type=type(d['post_exec'])) if 'cpu_reqs' in d: if isinstance(d['cpu_reqs'], dict): self._cpu_reqs = d['cpu_reqs'] else: raise TypeError(expected_type=dict, actual_type=type(d['cpu_reqs'])) if 'gpu_reqs' in d: if isinstance(d['gpu_reqs'], dict): self._gpu_reqs = d['gpu_reqs'] else: raise TypeError(expected_type=dict, actual_type=type(d['gpu_reqs'])) if 'lfs_per_process' in d: if d['lfs_per_process']: if isinstance(d['lfs_per_process'], int): self._lfs_per_process = d['lfs_per_process'] else: raise TypeError(expected_type=int, actual_type=type(d['lfs_per_process'])) if 'upload_input_data' in d: if isinstance(d['upload_input_data'], list): self._upload_input_data = d['upload_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['upload_input_data'])) if 'copy_input_data' in d: if isinstance(d['copy_input_data'], list): self._copy_input_data = d['copy_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['copy_input_data'])) if 'link_input_data' in d: if isinstance(d['link_input_data'], list): self._link_input_data = d['link_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['link_input_data'])) if 'move_input_data' in d: if isinstance(d['move_input_data'], list): self._move_input_data = d['move_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['move_input_data'])) if 'copy_output_data' in d: if isinstance(d['copy_output_data'], list): self._copy_output_data = d['copy_output_data'] else: raise TypeError(expected_type=list, actual_type=type(d['copy_output_data'])) if 'move_output_data' in d: if isinstance(d['move_output_data'], list): self._move_output_data = d['move_output_data'] else: raise TypeError(expected_type=list, actual_type=type(d['move_output_data'])) if 'download_output_data' in d: if isinstance(d['download_output_data'], list): self._download_output_data = d['download_output_data'] else: raise TypeError(expected_type=list, actual_type=type( d['download_output_data'])) if 'stdout' in d: if d['stdout']: if isinstance(d['stdout'], str) or isinstance(d['stdout'], unicode): self._stdout = d['stdout'] else: raise TypeError(expected_type=str, actual_type=type(d['stdout'])) if 'stderr' in d: if d['stderr']: if isinstance(d['stderr'], str) or isinstance(d['stderr'], unicode): self._stderr = d['stderr'] else: raise TypeError(expected_type=str, actual_type=type(d['stderr'])) if 'exit_code' in d: if d['exit_code']: if isinstance(d['exit_code'], int): self._exit_code = d['exit_code'] else: raise TypeError( entity='exit_code', expected_type=int, actual_type=type(d['exit_code'])) if 'path' in d: if d['path']: if isinstance(d['path'], str) or isinstance(d['path'], unicode): self._path = d['path'] else: raise TypeError(entity='path', expected_type=str, actual_type=type(d['path'])) if 'tag' in d: if d['tag']: if isinstance(d['tag'], str) or isinstance(d['tag'], unicode): self._tag = str(d['tag']) else: raise TypeError(expected_type=str, actual_type=type(d['tag'])) if 'parent_stage' in d: if isinstance(d['parent_stage'], dict): self._p_stage = d['parent_stage'] else: raise TypeError( entity='parent_stage', expected_type=dict, actual_type=type(d['parent_stage'])) if 'parent_pipeline' in d: if isinstance(d['parent_pipeline'], dict): self._p_pipeline = d['parent_pipeline'] else: raise TypeError(entity='parent_pipeline', expected_type=dict, actual_type=type( d['parent_pipeline']))
python
def from_dict(self, d): """ Create a Task from a dictionary. The change is in inplace. :argument: python dictionary :return: None """ if 'uid' in d: if d['uid']: self._uid = d['uid'] if 'name' in d: if d['name']: self._name = d['name'] if 'state' in d: if isinstance(d['state'], str) or isinstance(d['state'], unicode): self._state = d['state'] else: raise TypeError(entity='state', expected_type=str, actual_type=type(d['state'])) else: self._state = states.INITIAL if 'state_history' in d: if isinstance(d['state_history'], list): self._state_history = d['state_history'] else: raise TypeError(entity='state_history', expected_type=list, actual_type=type( d['state_history'])) if 'pre_exec' in d: if isinstance(d['pre_exec'], list): self._pre_exec = d['pre_exec'] else: raise TypeError(expected_type=list, actual_type=type(d['pre_exec'])) if 'executable' in d: if isinstance(d['executable'], str) or isinstance(d['executable'], unicode): self._executable = d['executable'] else: raise TypeError(expected_type=str, actual_type=type(d['executable'])) if 'arguments' in d: if isinstance(d['arguments'], list): self._arguments = d['arguments'] else: raise TypeError(expected_type=list, actual_type=type(d['arguments'])) if 'post_exec' in d: if isinstance(d['post_exec'], list): self._post_exec = d['post_exec'] else: raise TypeError(expected_type=list, actual_type=type(d['post_exec'])) if 'cpu_reqs' in d: if isinstance(d['cpu_reqs'], dict): self._cpu_reqs = d['cpu_reqs'] else: raise TypeError(expected_type=dict, actual_type=type(d['cpu_reqs'])) if 'gpu_reqs' in d: if isinstance(d['gpu_reqs'], dict): self._gpu_reqs = d['gpu_reqs'] else: raise TypeError(expected_type=dict, actual_type=type(d['gpu_reqs'])) if 'lfs_per_process' in d: if d['lfs_per_process']: if isinstance(d['lfs_per_process'], int): self._lfs_per_process = d['lfs_per_process'] else: raise TypeError(expected_type=int, actual_type=type(d['lfs_per_process'])) if 'upload_input_data' in d: if isinstance(d['upload_input_data'], list): self._upload_input_data = d['upload_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['upload_input_data'])) if 'copy_input_data' in d: if isinstance(d['copy_input_data'], list): self._copy_input_data = d['copy_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['copy_input_data'])) if 'link_input_data' in d: if isinstance(d['link_input_data'], list): self._link_input_data = d['link_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['link_input_data'])) if 'move_input_data' in d: if isinstance(d['move_input_data'], list): self._move_input_data = d['move_input_data'] else: raise TypeError(expected_type=list, actual_type=type(d['move_input_data'])) if 'copy_output_data' in d: if isinstance(d['copy_output_data'], list): self._copy_output_data = d['copy_output_data'] else: raise TypeError(expected_type=list, actual_type=type(d['copy_output_data'])) if 'move_output_data' in d: if isinstance(d['move_output_data'], list): self._move_output_data = d['move_output_data'] else: raise TypeError(expected_type=list, actual_type=type(d['move_output_data'])) if 'download_output_data' in d: if isinstance(d['download_output_data'], list): self._download_output_data = d['download_output_data'] else: raise TypeError(expected_type=list, actual_type=type( d['download_output_data'])) if 'stdout' in d: if d['stdout']: if isinstance(d['stdout'], str) or isinstance(d['stdout'], unicode): self._stdout = d['stdout'] else: raise TypeError(expected_type=str, actual_type=type(d['stdout'])) if 'stderr' in d: if d['stderr']: if isinstance(d['stderr'], str) or isinstance(d['stderr'], unicode): self._stderr = d['stderr'] else: raise TypeError(expected_type=str, actual_type=type(d['stderr'])) if 'exit_code' in d: if d['exit_code']: if isinstance(d['exit_code'], int): self._exit_code = d['exit_code'] else: raise TypeError( entity='exit_code', expected_type=int, actual_type=type(d['exit_code'])) if 'path' in d: if d['path']: if isinstance(d['path'], str) or isinstance(d['path'], unicode): self._path = d['path'] else: raise TypeError(entity='path', expected_type=str, actual_type=type(d['path'])) if 'tag' in d: if d['tag']: if isinstance(d['tag'], str) or isinstance(d['tag'], unicode): self._tag = str(d['tag']) else: raise TypeError(expected_type=str, actual_type=type(d['tag'])) if 'parent_stage' in d: if isinstance(d['parent_stage'], dict): self._p_stage = d['parent_stage'] else: raise TypeError( entity='parent_stage', expected_type=dict, actual_type=type(d['parent_stage'])) if 'parent_pipeline' in d: if isinstance(d['parent_pipeline'], dict): self._p_pipeline = d['parent_pipeline'] else: raise TypeError(entity='parent_pipeline', expected_type=dict, actual_type=type( d['parent_pipeline']))
[ "def", "from_dict", "(", "self", ",", "d", ")", ":", "if", "'uid'", "in", "d", ":", "if", "d", "[", "'uid'", "]", ":", "self", ".", "_uid", "=", "d", "[", "'uid'", "]", "if", "'name'", "in", "d", ":", "if", "d", "[", "'name'", "]", ":", "self", ".", "_name", "=", "d", "[", "'name'", "]", "if", "'state'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'state'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'state'", "]", ",", "unicode", ")", ":", "self", ".", "_state", "=", "d", "[", "'state'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'state'", ",", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'state'", "]", ")", ")", "else", ":", "self", ".", "_state", "=", "states", ".", "INITIAL", "if", "'state_history'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'state_history'", "]", ",", "list", ")", ":", "self", ".", "_state_history", "=", "d", "[", "'state_history'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'state_history'", ",", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'state_history'", "]", ")", ")", "if", "'pre_exec'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'pre_exec'", "]", ",", "list", ")", ":", "self", ".", "_pre_exec", "=", "d", "[", "'pre_exec'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'pre_exec'", "]", ")", ")", "if", "'executable'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'executable'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'executable'", "]", ",", "unicode", ")", ":", "self", ".", "_executable", "=", "d", "[", "'executable'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'executable'", "]", ")", ")", "if", "'arguments'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'arguments'", "]", ",", "list", ")", ":", "self", ".", "_arguments", "=", "d", "[", "'arguments'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'arguments'", "]", ")", ")", "if", "'post_exec'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'post_exec'", "]", ",", "list", ")", ":", "self", ".", "_post_exec", "=", "d", "[", "'post_exec'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'post_exec'", "]", ")", ")", "if", "'cpu_reqs'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'cpu_reqs'", "]", ",", "dict", ")", ":", "self", ".", "_cpu_reqs", "=", "d", "[", "'cpu_reqs'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "dict", ",", "actual_type", "=", "type", "(", "d", "[", "'cpu_reqs'", "]", ")", ")", "if", "'gpu_reqs'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'gpu_reqs'", "]", ",", "dict", ")", ":", "self", ".", "_gpu_reqs", "=", "d", "[", "'gpu_reqs'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "dict", ",", "actual_type", "=", "type", "(", "d", "[", "'gpu_reqs'", "]", ")", ")", "if", "'lfs_per_process'", "in", "d", ":", "if", "d", "[", "'lfs_per_process'", "]", ":", "if", "isinstance", "(", "d", "[", "'lfs_per_process'", "]", ",", "int", ")", ":", "self", ".", "_lfs_per_process", "=", "d", "[", "'lfs_per_process'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "int", ",", "actual_type", "=", "type", "(", "d", "[", "'lfs_per_process'", "]", ")", ")", "if", "'upload_input_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'upload_input_data'", "]", ",", "list", ")", ":", "self", ".", "_upload_input_data", "=", "d", "[", "'upload_input_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'upload_input_data'", "]", ")", ")", "if", "'copy_input_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'copy_input_data'", "]", ",", "list", ")", ":", "self", ".", "_copy_input_data", "=", "d", "[", "'copy_input_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'copy_input_data'", "]", ")", ")", "if", "'link_input_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'link_input_data'", "]", ",", "list", ")", ":", "self", ".", "_link_input_data", "=", "d", "[", "'link_input_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'link_input_data'", "]", ")", ")", "if", "'move_input_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'move_input_data'", "]", ",", "list", ")", ":", "self", ".", "_move_input_data", "=", "d", "[", "'move_input_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'move_input_data'", "]", ")", ")", "if", "'copy_output_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'copy_output_data'", "]", ",", "list", ")", ":", "self", ".", "_copy_output_data", "=", "d", "[", "'copy_output_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'copy_output_data'", "]", ")", ")", "if", "'move_output_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'move_output_data'", "]", ",", "list", ")", ":", "self", ".", "_move_output_data", "=", "d", "[", "'move_output_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'move_output_data'", "]", ")", ")", "if", "'download_output_data'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'download_output_data'", "]", ",", "list", ")", ":", "self", ".", "_download_output_data", "=", "d", "[", "'download_output_data'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "list", ",", "actual_type", "=", "type", "(", "d", "[", "'download_output_data'", "]", ")", ")", "if", "'stdout'", "in", "d", ":", "if", "d", "[", "'stdout'", "]", ":", "if", "isinstance", "(", "d", "[", "'stdout'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'stdout'", "]", ",", "unicode", ")", ":", "self", ".", "_stdout", "=", "d", "[", "'stdout'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'stdout'", "]", ")", ")", "if", "'stderr'", "in", "d", ":", "if", "d", "[", "'stderr'", "]", ":", "if", "isinstance", "(", "d", "[", "'stderr'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'stderr'", "]", ",", "unicode", ")", ":", "self", ".", "_stderr", "=", "d", "[", "'stderr'", "]", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'stderr'", "]", ")", ")", "if", "'exit_code'", "in", "d", ":", "if", "d", "[", "'exit_code'", "]", ":", "if", "isinstance", "(", "d", "[", "'exit_code'", "]", ",", "int", ")", ":", "self", ".", "_exit_code", "=", "d", "[", "'exit_code'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'exit_code'", ",", "expected_type", "=", "int", ",", "actual_type", "=", "type", "(", "d", "[", "'exit_code'", "]", ")", ")", "if", "'path'", "in", "d", ":", "if", "d", "[", "'path'", "]", ":", "if", "isinstance", "(", "d", "[", "'path'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'path'", "]", ",", "unicode", ")", ":", "self", ".", "_path", "=", "d", "[", "'path'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'path'", ",", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'path'", "]", ")", ")", "if", "'tag'", "in", "d", ":", "if", "d", "[", "'tag'", "]", ":", "if", "isinstance", "(", "d", "[", "'tag'", "]", ",", "str", ")", "or", "isinstance", "(", "d", "[", "'tag'", "]", ",", "unicode", ")", ":", "self", ".", "_tag", "=", "str", "(", "d", "[", "'tag'", "]", ")", "else", ":", "raise", "TypeError", "(", "expected_type", "=", "str", ",", "actual_type", "=", "type", "(", "d", "[", "'tag'", "]", ")", ")", "if", "'parent_stage'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'parent_stage'", "]", ",", "dict", ")", ":", "self", ".", "_p_stage", "=", "d", "[", "'parent_stage'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'parent_stage'", ",", "expected_type", "=", "dict", ",", "actual_type", "=", "type", "(", "d", "[", "'parent_stage'", "]", ")", ")", "if", "'parent_pipeline'", "in", "d", ":", "if", "isinstance", "(", "d", "[", "'parent_pipeline'", "]", ",", "dict", ")", ":", "self", ".", "_p_pipeline", "=", "d", "[", "'parent_pipeline'", "]", "else", ":", "raise", "TypeError", "(", "entity", "=", "'parent_pipeline'", ",", "expected_type", "=", "dict", ",", "actual_type", "=", "type", "(", "d", "[", "'parent_pipeline'", "]", ")", ")" ]
Create a Task from a dictionary. The change is in inplace. :argument: python dictionary :return: None
[ "Create", "a", "Task", "from", "a", "dictionary", ".", "The", "change", "is", "in", "inplace", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/task/task.py#L766-L948
radical-cybertools/radical.entk
src/radical/entk/task/task.py
Task._assign_uid
def _assign_uid(self, sid): """ Purpose: Assign a uid to the current object based on the sid passed """ self._uid = ru.generate_id( 'task.%(item_counter)04d', ru.ID_CUSTOM, namespace=sid)
python
def _assign_uid(self, sid): """ Purpose: Assign a uid to the current object based on the sid passed """ self._uid = ru.generate_id( 'task.%(item_counter)04d', ru.ID_CUSTOM, namespace=sid)
[ "def", "_assign_uid", "(", "self", ",", "sid", ")", ":", "self", ".", "_uid", "=", "ru", ".", "generate_id", "(", "'task.%(item_counter)04d'", ",", "ru", ".", "ID_CUSTOM", ",", "namespace", "=", "sid", ")" ]
Purpose: Assign a uid to the current object based on the sid passed
[ "Purpose", ":", "Assign", "a", "uid", "to", "the", "current", "object", "based", "on", "the", "sid", "passed" ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/task/task.py#L954-L959
radical-cybertools/radical.entk
src/radical/entk/task/task.py
Task._validate
def _validate(self): """ Purpose: Validate that the state of the task is 'DESCRIBED' and that an executable has been specified for the task. """ if self._state is not states.INITIAL: raise ValueError(obj=self._uid, attribute='state', expected_value=states.INITIAL, actual_value=self._state) if not self._executable: raise MissingError(obj=self._uid, missing_attribute='executable')
python
def _validate(self): """ Purpose: Validate that the state of the task is 'DESCRIBED' and that an executable has been specified for the task. """ if self._state is not states.INITIAL: raise ValueError(obj=self._uid, attribute='state', expected_value=states.INITIAL, actual_value=self._state) if not self._executable: raise MissingError(obj=self._uid, missing_attribute='executable')
[ "def", "_validate", "(", "self", ")", ":", "if", "self", ".", "_state", "is", "not", "states", ".", "INITIAL", ":", "raise", "ValueError", "(", "obj", "=", "self", ".", "_uid", ",", "attribute", "=", "'state'", ",", "expected_value", "=", "states", ".", "INITIAL", ",", "actual_value", "=", "self", ".", "_state", ")", "if", "not", "self", ".", "_executable", ":", "raise", "MissingError", "(", "obj", "=", "self", ".", "_uid", ",", "missing_attribute", "=", "'executable'", ")" ]
Purpose: Validate that the state of the task is 'DESCRIBED' and that an executable has been specified for the task.
[ "Purpose", ":", "Validate", "that", "the", "state", "of", "the", "task", "is", "DESCRIBED", "and", "that", "an", "executable", "has", "been", "specified", "for", "the", "task", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/task/task.py#L961-L975
radical-cybertools/radical.entk
src/radical/entk/execman/rp/task_manager.py
TaskManager._process_tasks
def _process_tasks(self, task_queue, rmgr, logger, mq_hostname, port, local_prof, sid): ''' **Purpose**: The new thread that gets spawned by the main tmgr process invokes this function. This function receives tasks from 'task_queue' and submits them to the RADICAL Pilot RTS. ''' placeholder_dict = dict() def load_placeholder(task, rts_uid): parent_pipeline = str(task.parent_pipeline['name']) parent_stage = str(task.parent_stage['name']) if parent_pipeline not in placeholder_dict: placeholder_dict[parent_pipeline] = dict() if parent_stage not in placeholder_dict[parent_pipeline]: placeholder_dict[parent_pipeline][parent_stage] = dict() if None not in [parent_pipeline, parent_stage, task.name]: placeholder_dict[parent_pipeline][parent_stage][str(task.name)] = {'path': str(task.path), 'rts_uid': rts_uid} def unit_state_cb(unit, state): try: logger.debug('Unit %s in state %s' % (unit.uid, unit.state)) if unit.state in rp.FINAL: # Acquire a connection+channel to the rmq server mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() task = None task = create_task_from_cu(unit, local_prof) transition(obj=task, obj_type='Task', new_state=states.COMPLETED, channel=mq_channel, queue='%s-cb-to-sync' % sid, profiler=local_prof, logger=logger) load_placeholder(task, unit.uid) task_as_dict = json.dumps(task.to_dict()) mq_channel.basic_publish(exchange='', routing_key='%s-completedq-1' % sid, body=task_as_dict # properties=pika.BasicProperties( # make message persistent # delivery_mode = 2, # ) ) logger.info('Pushed task %s with state %s to completed queue %s-completedq-1' % (task.uid, task.state, sid)) mq_connection.close() except KeyboardInterrupt: logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit callback thread gracefully...') raise KeyboardInterrupt except Exception, ex: logger.exception('Error in RP callback thread: %s' % ex) umgr = rp.UnitManager(session=rmgr._session) umgr.add_pilots(rmgr.pilot) umgr.register_callback(unit_state_cb) try: while not self._tmgr_terminate.is_set(): body = None try: body = task_queue.get(block=True, timeout=10) except Queue.Empty: # Ignore empty exception, we don't always have new tasks to run pass if body: task_queue.task_done() bulk_tasks = list() bulk_cuds = list() for task in body: t = Task() t.from_dict(task) bulk_tasks.append(t) bulk_cuds.append(create_cud_from_task( t, placeholder_dict, local_prof)) mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() transition(obj=t, obj_type='Task', new_state=states.SUBMITTING, channel=mq_channel, queue='%s-tmgr-to-sync' % sid, profiler=local_prof, logger=logger) mq_connection.close() umgr.submit_units(bulk_cuds) for task in bulk_tasks: mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() transition(obj=task, obj_type='Task', new_state=states.SUBMITTED, channel=mq_channel, queue='%s-tmgr-to-sync' % sid, profiler=local_prof, logger=logger) mq_connection.close() except KeyboardInterrupt as ex: logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel task processor gracefully...') except Exception as ex: logger.exception('%s failed with %s'%(self._uid, ex)) raise EnTKError(ex)
python
def _process_tasks(self, task_queue, rmgr, logger, mq_hostname, port, local_prof, sid): ''' **Purpose**: The new thread that gets spawned by the main tmgr process invokes this function. This function receives tasks from 'task_queue' and submits them to the RADICAL Pilot RTS. ''' placeholder_dict = dict() def load_placeholder(task, rts_uid): parent_pipeline = str(task.parent_pipeline['name']) parent_stage = str(task.parent_stage['name']) if parent_pipeline not in placeholder_dict: placeholder_dict[parent_pipeline] = dict() if parent_stage not in placeholder_dict[parent_pipeline]: placeholder_dict[parent_pipeline][parent_stage] = dict() if None not in [parent_pipeline, parent_stage, task.name]: placeholder_dict[parent_pipeline][parent_stage][str(task.name)] = {'path': str(task.path), 'rts_uid': rts_uid} def unit_state_cb(unit, state): try: logger.debug('Unit %s in state %s' % (unit.uid, unit.state)) if unit.state in rp.FINAL: # Acquire a connection+channel to the rmq server mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() task = None task = create_task_from_cu(unit, local_prof) transition(obj=task, obj_type='Task', new_state=states.COMPLETED, channel=mq_channel, queue='%s-cb-to-sync' % sid, profiler=local_prof, logger=logger) load_placeholder(task, unit.uid) task_as_dict = json.dumps(task.to_dict()) mq_channel.basic_publish(exchange='', routing_key='%s-completedq-1' % sid, body=task_as_dict # properties=pika.BasicProperties( # make message persistent # delivery_mode = 2, # ) ) logger.info('Pushed task %s with state %s to completed queue %s-completedq-1' % (task.uid, task.state, sid)) mq_connection.close() except KeyboardInterrupt: logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to exit callback thread gracefully...') raise KeyboardInterrupt except Exception, ex: logger.exception('Error in RP callback thread: %s' % ex) umgr = rp.UnitManager(session=rmgr._session) umgr.add_pilots(rmgr.pilot) umgr.register_callback(unit_state_cb) try: while not self._tmgr_terminate.is_set(): body = None try: body = task_queue.get(block=True, timeout=10) except Queue.Empty: # Ignore empty exception, we don't always have new tasks to run pass if body: task_queue.task_done() bulk_tasks = list() bulk_cuds = list() for task in body: t = Task() t.from_dict(task) bulk_tasks.append(t) bulk_cuds.append(create_cud_from_task( t, placeholder_dict, local_prof)) mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() transition(obj=t, obj_type='Task', new_state=states.SUBMITTING, channel=mq_channel, queue='%s-tmgr-to-sync' % sid, profiler=local_prof, logger=logger) mq_connection.close() umgr.submit_units(bulk_cuds) for task in bulk_tasks: mq_connection = pika.BlockingConnection(pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() transition(obj=task, obj_type='Task', new_state=states.SUBMITTED, channel=mq_channel, queue='%s-tmgr-to-sync' % sid, profiler=local_prof, logger=logger) mq_connection.close() except KeyboardInterrupt as ex: logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel task processor gracefully...') except Exception as ex: logger.exception('%s failed with %s'%(self._uid, ex)) raise EnTKError(ex)
[ "def", "_process_tasks", "(", "self", ",", "task_queue", ",", "rmgr", ",", "logger", ",", "mq_hostname", ",", "port", ",", "local_prof", ",", "sid", ")", ":", "placeholder_dict", "=", "dict", "(", ")", "def", "load_placeholder", "(", "task", ",", "rts_uid", ")", ":", "parent_pipeline", "=", "str", "(", "task", ".", "parent_pipeline", "[", "'name'", "]", ")", "parent_stage", "=", "str", "(", "task", ".", "parent_stage", "[", "'name'", "]", ")", "if", "parent_pipeline", "not", "in", "placeholder_dict", ":", "placeholder_dict", "[", "parent_pipeline", "]", "=", "dict", "(", ")", "if", "parent_stage", "not", "in", "placeholder_dict", "[", "parent_pipeline", "]", ":", "placeholder_dict", "[", "parent_pipeline", "]", "[", "parent_stage", "]", "=", "dict", "(", ")", "if", "None", "not", "in", "[", "parent_pipeline", ",", "parent_stage", ",", "task", ".", "name", "]", ":", "placeholder_dict", "[", "parent_pipeline", "]", "[", "parent_stage", "]", "[", "str", "(", "task", ".", "name", ")", "]", "=", "{", "'path'", ":", "str", "(", "task", ".", "path", ")", ",", "'rts_uid'", ":", "rts_uid", "}", "def", "unit_state_cb", "(", "unit", ",", "state", ")", ":", "try", ":", "logger", ".", "debug", "(", "'Unit %s in state %s'", "%", "(", "unit", ".", "uid", ",", "unit", ".", "state", ")", ")", "if", "unit", ".", "state", "in", "rp", ".", "FINAL", ":", "# Acquire a connection+channel to the rmq server", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "mq_hostname", ",", "port", "=", "port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "task", "=", "None", "task", "=", "create_task_from_cu", "(", "unit", ",", "local_prof", ")", "transition", "(", "obj", "=", "task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "COMPLETED", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-cb-to-sync'", "%", "sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "logger", ")", "load_placeholder", "(", "task", ",", "unit", ".", "uid", ")", "task_as_dict", "=", "json", ".", "dumps", "(", "task", ".", "to_dict", "(", ")", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "'%s-completedq-1'", "%", "sid", ",", "body", "=", "task_as_dict", "# properties=pika.BasicProperties(", "# make message persistent", "# delivery_mode = 2,", "# )", ")", "logger", ".", "info", "(", "'Pushed task %s with state %s to completed queue %s-completedq-1'", "%", "(", "task", ".", "uid", ",", "task", ".", "state", ",", "sid", ")", ")", "mq_connection", ".", "close", "(", ")", "except", "KeyboardInterrupt", ":", "logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to exit callback thread gracefully...'", ")", "raise", "KeyboardInterrupt", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Error in RP callback thread: %s'", "%", "ex", ")", "umgr", "=", "rp", ".", "UnitManager", "(", "session", "=", "rmgr", ".", "_session", ")", "umgr", ".", "add_pilots", "(", "rmgr", ".", "pilot", ")", "umgr", ".", "register_callback", "(", "unit_state_cb", ")", "try", ":", "while", "not", "self", ".", "_tmgr_terminate", ".", "is_set", "(", ")", ":", "body", "=", "None", "try", ":", "body", "=", "task_queue", ".", "get", "(", "block", "=", "True", ",", "timeout", "=", "10", ")", "except", "Queue", ".", "Empty", ":", "# Ignore empty exception, we don't always have new tasks to run", "pass", "if", "body", ":", "task_queue", ".", "task_done", "(", ")", "bulk_tasks", "=", "list", "(", ")", "bulk_cuds", "=", "list", "(", ")", "for", "task", "in", "body", ":", "t", "=", "Task", "(", ")", "t", ".", "from_dict", "(", "task", ")", "bulk_tasks", ".", "append", "(", "t", ")", "bulk_cuds", ".", "append", "(", "create_cud_from_task", "(", "t", ",", "placeholder_dict", ",", "local_prof", ")", ")", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "mq_hostname", ",", "port", "=", "port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "transition", "(", "obj", "=", "t", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "SUBMITTING", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-tmgr-to-sync'", "%", "sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "logger", ")", "mq_connection", ".", "close", "(", ")", "umgr", ".", "submit_units", "(", "bulk_cuds", ")", "for", "task", "in", "bulk_tasks", ":", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "mq_hostname", ",", "port", "=", "port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "transition", "(", "obj", "=", "task", ",", "obj_type", "=", "'Task'", ",", "new_state", "=", "states", ".", "SUBMITTED", ",", "channel", "=", "mq_channel", ",", "queue", "=", "'%s-tmgr-to-sync'", "%", "sid", ",", "profiler", "=", "local_prof", ",", "logger", "=", "logger", ")", "mq_connection", ".", "close", "(", ")", "except", "KeyboardInterrupt", "as", "ex", ":", "logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to cancel task processor gracefully...'", ")", "except", "Exception", "as", "ex", ":", "logger", ".", "exception", "(", "'%s failed with %s'", "%", "(", "self", ".", "_uid", ",", "ex", ")", ")", "raise", "EnTKError", "(", "ex", ")" ]
**Purpose**: The new thread that gets spawned by the main tmgr process invokes this function. This function receives tasks from 'task_queue' and submits them to the RADICAL Pilot RTS.
[ "**", "Purpose", "**", ":", "The", "new", "thread", "that", "gets", "spawned", "by", "the", "main", "tmgr", "process", "invokes", "this", "function", ".", "This", "function", "receives", "tasks", "from", "task_queue", "and", "submits", "them", "to", "the", "RADICAL", "Pilot", "RTS", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/rp/task_manager.py#L187-L330
google/dotty
efilter/transforms/infer_type.py
infer_type
def infer_type(expr, scope): """Try to infer the type of x[y] if y is a known value (literal).""" # Do we know what the key even is? if isinstance(expr.key, ast.Literal): key = expr.key.value else: return protocol.AnyType container_type = infer_type(expr.value, scope) try: # Associative types are not subject to scoping rules so we can just # reflect using IAssociative. return associative.reflect(container_type, key) or protocol.AnyType except NotImplementedError: return protocol.AnyType
python
def infer_type(expr, scope): """Try to infer the type of x[y] if y is a known value (literal).""" # Do we know what the key even is? if isinstance(expr.key, ast.Literal): key = expr.key.value else: return protocol.AnyType container_type = infer_type(expr.value, scope) try: # Associative types are not subject to scoping rules so we can just # reflect using IAssociative. return associative.reflect(container_type, key) or protocol.AnyType except NotImplementedError: return protocol.AnyType
[ "def", "infer_type", "(", "expr", ",", "scope", ")", ":", "# Do we know what the key even is?", "if", "isinstance", "(", "expr", ".", "key", ",", "ast", ".", "Literal", ")", ":", "key", "=", "expr", ".", "key", ".", "value", "else", ":", "return", "protocol", ".", "AnyType", "container_type", "=", "infer_type", "(", "expr", ".", "value", ",", "scope", ")", "try", ":", "# Associative types are not subject to scoping rules so we can just", "# reflect using IAssociative.", "return", "associative", ".", "reflect", "(", "container_type", ",", "key", ")", "or", "protocol", ".", "AnyType", "except", "NotImplementedError", ":", "return", "protocol", ".", "AnyType" ]
Try to infer the type of x[y] if y is a known value (literal).
[ "Try", "to", "infer", "the", "type", "of", "x", "[", "y", "]", "if", "y", "is", "a", "known", "value", "(", "literal", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/infer_type.py#L108-L123
google/dotty
efilter/transforms/infer_type.py
infer_type
def infer_type(expr, scope): """Try to infer the type of x.y if y is a known value (literal).""" # Do we know what the member is? if isinstance(expr.member, ast.Literal): member = expr.member.value else: return protocol.AnyType container_type = infer_type(expr.obj, scope) try: # We are not using lexical scope here on purpose - we want to see what # the type of the member is only on the container_type. return structured.reflect(container_type, member) or protocol.AnyType except NotImplementedError: return protocol.AnyType
python
def infer_type(expr, scope): """Try to infer the type of x.y if y is a known value (literal).""" # Do we know what the member is? if isinstance(expr.member, ast.Literal): member = expr.member.value else: return protocol.AnyType container_type = infer_type(expr.obj, scope) try: # We are not using lexical scope here on purpose - we want to see what # the type of the member is only on the container_type. return structured.reflect(container_type, member) or protocol.AnyType except NotImplementedError: return protocol.AnyType
[ "def", "infer_type", "(", "expr", ",", "scope", ")", ":", "# Do we know what the member is?", "if", "isinstance", "(", "expr", ".", "member", ",", "ast", ".", "Literal", ")", ":", "member", "=", "expr", ".", "member", ".", "value", "else", ":", "return", "protocol", ".", "AnyType", "container_type", "=", "infer_type", "(", "expr", ".", "obj", ",", "scope", ")", "try", ":", "# We are not using lexical scope here on purpose - we want to see what", "# the type of the member is only on the container_type.", "return", "structured", ".", "reflect", "(", "container_type", ",", "member", ")", "or", "protocol", ".", "AnyType", "except", "NotImplementedError", ":", "return", "protocol", ".", "AnyType" ]
Try to infer the type of x.y if y is a known value (literal).
[ "Try", "to", "infer", "the", "type", "of", "x", ".", "y", "if", "y", "is", "a", "known", "value", "(", "literal", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/transforms/infer_type.py#L127-L142
radical-cybertools/radical.entk
src/radical/entk/execman/mock/task_manager.py
TaskManager._tmgr
def _tmgr(self, uid, rmgr, logger, mq_hostname, port, pending_queue, completed_queue): """ **Purpose**: Method to be run by the tmgr process. This method receives a Task from the pending_queue and submits it to the RTS. Currently, it also converts Tasks into CUDs and CUs into (partially described) Tasks. This conversion is necessary since the current RTS is RADICAL Pilot. Once Tasks are recovered from a CU, they are then pushed to the completed_queue. At all state transititons, they are synced (blocking) with the AppManager in the master process. In addition the tmgr also receives heartbeat 'request' msgs from the heartbeat-req queue. It responds with a 'response' message to the 'heartbeart-res' queue. **Details**: The AppManager can re-invoke the tmgr process with this function if the execution of the workflow is still incomplete. There is also population of a dictionary, placeholder_dict, which stores the path of each of the tasks on the remote machine. """ try: def heartbeat_response(mq_channel): try: # Get request from heartbeat-req for heartbeat response hb_method_frame, hb_props, hb_body = mq_channel.basic_get( queue=self._hb_request_q) if hb_body: logger.info('Received heartbeat request') mq_channel.basic_publish(exchange='', routing_key=self._hb_response_q, properties=pika.BasicProperties( correlation_id=hb_props.correlation_id), body='response') logger.info('Sent heartbeat response') mq_channel.basic_ack( delivery_tag=hb_method_frame.delivery_tag) except Exception, ex: logger.exception( 'Failed to respond to heartbeat request, error: %s' % ex) raise local_prof = ru.Profiler( name='radical.entk.%s' % self._uid + '-proc', path=self._path) local_prof.prof('tmgr process started', uid=self._uid) logger.info('Task Manager process started') # Thread should run till terminate condtion is encountered mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() # Queue for communication between threads of this process task_queue = Queue.Queue() # Start second thread to receive tasks and push to RTS self._rts_runner = threading.Thread(target=self._process_tasks, args=(task_queue, rmgr, logger, mq_hostname, port, local_prof, self._sid)) self._rts_runner.start() local_prof.prof('tmgr infrastructure setup done', uid=uid) last = time.time() while not self._tmgr_terminate.is_set(): try: method_frame, header_frame, body = mq_channel.basic_get( queue=pending_queue[0]) if body: body = json.loads(body) task_queue.put(body) mq_channel.basic_ack( delivery_tag=method_frame.delivery_tag) heartbeat_response(mq_channel) except Exception, ex: logger.exception('Error in task execution: %s' % ex) raise except KeyboardInterrupt: logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel tmgr process gracefully...') except Exception, ex: logger.exception('%s failed with %s'%(self._uid, ex)) raise EnTKError(ex) finally: local_prof.prof('terminating tmgr process', uid=uid) if self._rts_runner: self._rts_runner.join() mq_connection.close() local_prof.close()
python
def _tmgr(self, uid, rmgr, logger, mq_hostname, port, pending_queue, completed_queue): """ **Purpose**: Method to be run by the tmgr process. This method receives a Task from the pending_queue and submits it to the RTS. Currently, it also converts Tasks into CUDs and CUs into (partially described) Tasks. This conversion is necessary since the current RTS is RADICAL Pilot. Once Tasks are recovered from a CU, they are then pushed to the completed_queue. At all state transititons, they are synced (blocking) with the AppManager in the master process. In addition the tmgr also receives heartbeat 'request' msgs from the heartbeat-req queue. It responds with a 'response' message to the 'heartbeart-res' queue. **Details**: The AppManager can re-invoke the tmgr process with this function if the execution of the workflow is still incomplete. There is also population of a dictionary, placeholder_dict, which stores the path of each of the tasks on the remote machine. """ try: def heartbeat_response(mq_channel): try: # Get request from heartbeat-req for heartbeat response hb_method_frame, hb_props, hb_body = mq_channel.basic_get( queue=self._hb_request_q) if hb_body: logger.info('Received heartbeat request') mq_channel.basic_publish(exchange='', routing_key=self._hb_response_q, properties=pika.BasicProperties( correlation_id=hb_props.correlation_id), body='response') logger.info('Sent heartbeat response') mq_channel.basic_ack( delivery_tag=hb_method_frame.delivery_tag) except Exception, ex: logger.exception( 'Failed to respond to heartbeat request, error: %s' % ex) raise local_prof = ru.Profiler( name='radical.entk.%s' % self._uid + '-proc', path=self._path) local_prof.prof('tmgr process started', uid=self._uid) logger.info('Task Manager process started') # Thread should run till terminate condtion is encountered mq_connection = pika.BlockingConnection( pika.ConnectionParameters(host=mq_hostname, port=port)) mq_channel = mq_connection.channel() # Queue for communication between threads of this process task_queue = Queue.Queue() # Start second thread to receive tasks and push to RTS self._rts_runner = threading.Thread(target=self._process_tasks, args=(task_queue, rmgr, logger, mq_hostname, port, local_prof, self._sid)) self._rts_runner.start() local_prof.prof('tmgr infrastructure setup done', uid=uid) last = time.time() while not self._tmgr_terminate.is_set(): try: method_frame, header_frame, body = mq_channel.basic_get( queue=pending_queue[0]) if body: body = json.loads(body) task_queue.put(body) mq_channel.basic_ack( delivery_tag=method_frame.delivery_tag) heartbeat_response(mq_channel) except Exception, ex: logger.exception('Error in task execution: %s' % ex) raise except KeyboardInterrupt: logger.exception('Execution interrupted by user (you probably hit Ctrl+C), ' + 'trying to cancel tmgr process gracefully...') except Exception, ex: logger.exception('%s failed with %s'%(self._uid, ex)) raise EnTKError(ex) finally: local_prof.prof('terminating tmgr process', uid=uid) if self._rts_runner: self._rts_runner.join() mq_connection.close() local_prof.close()
[ "def", "_tmgr", "(", "self", ",", "uid", ",", "rmgr", ",", "logger", ",", "mq_hostname", ",", "port", ",", "pending_queue", ",", "completed_queue", ")", ":", "try", ":", "def", "heartbeat_response", "(", "mq_channel", ")", ":", "try", ":", "# Get request from heartbeat-req for heartbeat response", "hb_method_frame", ",", "hb_props", ",", "hb_body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "self", ".", "_hb_request_q", ")", "if", "hb_body", ":", "logger", ".", "info", "(", "'Received heartbeat request'", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "''", ",", "routing_key", "=", "self", ".", "_hb_response_q", ",", "properties", "=", "pika", ".", "BasicProperties", "(", "correlation_id", "=", "hb_props", ".", "correlation_id", ")", ",", "body", "=", "'response'", ")", "logger", ".", "info", "(", "'Sent heartbeat response'", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "hb_method_frame", ".", "delivery_tag", ")", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Failed to respond to heartbeat request, error: %s'", "%", "ex", ")", "raise", "local_prof", "=", "ru", ".", "Profiler", "(", "name", "=", "'radical.entk.%s'", "%", "self", ".", "_uid", "+", "'-proc'", ",", "path", "=", "self", ".", "_path", ")", "local_prof", ".", "prof", "(", "'tmgr process started'", ",", "uid", "=", "self", ".", "_uid", ")", "logger", ".", "info", "(", "'Task Manager process started'", ")", "# Thread should run till terminate condtion is encountered", "mq_connection", "=", "pika", ".", "BlockingConnection", "(", "pika", ".", "ConnectionParameters", "(", "host", "=", "mq_hostname", ",", "port", "=", "port", ")", ")", "mq_channel", "=", "mq_connection", ".", "channel", "(", ")", "# Queue for communication between threads of this process", "task_queue", "=", "Queue", ".", "Queue", "(", ")", "# Start second thread to receive tasks and push to RTS", "self", ".", "_rts_runner", "=", "threading", ".", "Thread", "(", "target", "=", "self", ".", "_process_tasks", ",", "args", "=", "(", "task_queue", ",", "rmgr", ",", "logger", ",", "mq_hostname", ",", "port", ",", "local_prof", ",", "self", ".", "_sid", ")", ")", "self", ".", "_rts_runner", ".", "start", "(", ")", "local_prof", ".", "prof", "(", "'tmgr infrastructure setup done'", ",", "uid", "=", "uid", ")", "last", "=", "time", ".", "time", "(", ")", "while", "not", "self", ".", "_tmgr_terminate", ".", "is_set", "(", ")", ":", "try", ":", "method_frame", ",", "header_frame", ",", "body", "=", "mq_channel", ".", "basic_get", "(", "queue", "=", "pending_queue", "[", "0", "]", ")", "if", "body", ":", "body", "=", "json", ".", "loads", "(", "body", ")", "task_queue", ".", "put", "(", "body", ")", "mq_channel", ".", "basic_ack", "(", "delivery_tag", "=", "method_frame", ".", "delivery_tag", ")", "heartbeat_response", "(", "mq_channel", ")", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'Error in task execution: %s'", "%", "ex", ")", "raise", "except", "KeyboardInterrupt", ":", "logger", ".", "exception", "(", "'Execution interrupted by user (you probably hit Ctrl+C), '", "+", "'trying to cancel tmgr process gracefully...'", ")", "except", "Exception", ",", "ex", ":", "logger", ".", "exception", "(", "'%s failed with %s'", "%", "(", "self", ".", "_uid", ",", "ex", ")", ")", "raise", "EnTKError", "(", "ex", ")", "finally", ":", "local_prof", ".", "prof", "(", "'terminating tmgr process'", ",", "uid", "=", "uid", ")", "if", "self", ".", "_rts_runner", ":", "self", ".", "_rts_runner", ".", "join", "(", ")", "mq_connection", ".", "close", "(", ")", "local_prof", ".", "close", "(", ")" ]
**Purpose**: Method to be run by the tmgr process. This method receives a Task from the pending_queue and submits it to the RTS. Currently, it also converts Tasks into CUDs and CUs into (partially described) Tasks. This conversion is necessary since the current RTS is RADICAL Pilot. Once Tasks are recovered from a CU, they are then pushed to the completed_queue. At all state transititons, they are synced (blocking) with the AppManager in the master process. In addition the tmgr also receives heartbeat 'request' msgs from the heartbeat-req queue. It responds with a 'response' message to the 'heartbeart-res' queue. **Details**: The AppManager can re-invoke the tmgr process with this function if the execution of the workflow is still incomplete. There is also population of a dictionary, placeholder_dict, which stores the path of each of the tasks on the remote machine.
[ "**", "Purpose", "**", ":", "Method", "to", "be", "run", "by", "the", "tmgr", "process", ".", "This", "method", "receives", "a", "Task", "from", "the", "pending_queue", "and", "submits", "it", "to", "the", "RTS", ".", "Currently", "it", "also", "converts", "Tasks", "into", "CUDs", "and", "CUs", "into", "(", "partially", "described", ")", "Tasks", ".", "This", "conversion", "is", "necessary", "since", "the", "current", "RTS", "is", "RADICAL", "Pilot", ".", "Once", "Tasks", "are", "recovered", "from", "a", "CU", "they", "are", "then", "pushed", "to", "the", "completed_queue", ".", "At", "all", "state", "transititons", "they", "are", "synced", "(", "blocking", ")", "with", "the", "AppManager", "in", "the", "master", "process", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/mock/task_manager.py#L59-L171
radical-cybertools/radical.entk
src/radical/entk/execman/mock/task_manager.py
TaskManager.start_manager
def start_manager(self): """ **Purpose**: Method to start the tmgr process. The tmgr function is not to be accessed directly. The function is started in a separate thread using this method. """ if not self._tmgr_process: try: self._prof.prof('creating tmgr process', uid=self._uid) self._tmgr_terminate = Event() self._tmgr_process = Process(target=self._tmgr, name='task-manager', args=( self._uid, self._rmgr, self._logger, self._mq_hostname, self._port, self._pending_queue, self._completed_queue) ) self._logger.info('Starting task manager process') self._prof.prof('starting tmgr process', uid=self._uid) self._tmgr_process.start() return True except Exception, ex: self._logger.exception('Task manager not started, error: %s' % ex) self.terminate_manager() raise else: self._logger.warn( 'tmgr process already running, but attempted to restart!')
python
def start_manager(self): """ **Purpose**: Method to start the tmgr process. The tmgr function is not to be accessed directly. The function is started in a separate thread using this method. """ if not self._tmgr_process: try: self._prof.prof('creating tmgr process', uid=self._uid) self._tmgr_terminate = Event() self._tmgr_process = Process(target=self._tmgr, name='task-manager', args=( self._uid, self._rmgr, self._logger, self._mq_hostname, self._port, self._pending_queue, self._completed_queue) ) self._logger.info('Starting task manager process') self._prof.prof('starting tmgr process', uid=self._uid) self._tmgr_process.start() return True except Exception, ex: self._logger.exception('Task manager not started, error: %s' % ex) self.terminate_manager() raise else: self._logger.warn( 'tmgr process already running, but attempted to restart!')
[ "def", "start_manager", "(", "self", ")", ":", "if", "not", "self", ".", "_tmgr_process", ":", "try", ":", "self", ".", "_prof", ".", "prof", "(", "'creating tmgr process'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_tmgr_terminate", "=", "Event", "(", ")", "self", ".", "_tmgr_process", "=", "Process", "(", "target", "=", "self", ".", "_tmgr", ",", "name", "=", "'task-manager'", ",", "args", "=", "(", "self", ".", "_uid", ",", "self", ".", "_rmgr", ",", "self", ".", "_logger", ",", "self", ".", "_mq_hostname", ",", "self", ".", "_port", ",", "self", ".", "_pending_queue", ",", "self", ".", "_completed_queue", ")", ")", "self", ".", "_logger", ".", "info", "(", "'Starting task manager process'", ")", "self", ".", "_prof", ".", "prof", "(", "'starting tmgr process'", ",", "uid", "=", "self", ".", "_uid", ")", "self", ".", "_tmgr_process", ".", "start", "(", ")", "return", "True", "except", "Exception", ",", "ex", ":", "self", ".", "_logger", ".", "exception", "(", "'Task manager not started, error: %s'", "%", "ex", ")", "self", ".", "terminate_manager", "(", ")", "raise", "else", ":", "self", ".", "_logger", ".", "warn", "(", "'tmgr process already running, but attempted to restart!'", ")" ]
**Purpose**: Method to start the tmgr process. The tmgr function is not to be accessed directly. The function is started in a separate thread using this method.
[ "**", "Purpose", "**", ":", "Method", "to", "start", "the", "tmgr", "process", ".", "The", "tmgr", "function", "is", "not", "to", "be", "accessed", "directly", ".", "The", "function", "is", "started", "in", "a", "separate", "thread", "using", "this", "method", "." ]
train
https://github.com/radical-cybertools/radical.entk/blob/945f6c93c9a62db90ad191b306418d5c1cdd9d24/src/radical/entk/execman/mock/task_manager.py#L277-L317
google/dotty
efilter/parsers/common/grammar.py
keyword
def keyword(tokens, expected): """Case-insensitive keyword match.""" try: token = next(iter(tokens)) except StopIteration: return if token and token.name == "symbol" and token.value.lower() == expected: return TokenMatch(None, token.value, (token,))
python
def keyword(tokens, expected): """Case-insensitive keyword match.""" try: token = next(iter(tokens)) except StopIteration: return if token and token.name == "symbol" and token.value.lower() == expected: return TokenMatch(None, token.value, (token,))
[ "def", "keyword", "(", "tokens", ",", "expected", ")", ":", "try", ":", "token", "=", "next", "(", "iter", "(", "tokens", ")", ")", "except", "StopIteration", ":", "return", "if", "token", "and", "token", ".", "name", "==", "\"symbol\"", "and", "token", ".", "value", ".", "lower", "(", ")", "==", "expected", ":", "return", "TokenMatch", "(", "None", ",", "token", ".", "value", ",", "(", "token", ",", ")", ")" ]
Case-insensitive keyword match.
[ "Case", "-", "insensitive", "keyword", "match", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L234-L242
google/dotty
efilter/parsers/common/grammar.py
multi_keyword
def multi_keyword(tokens, keyword_parts): """Match a case-insensitive keyword consisting of multiple tokens.""" tokens = iter(tokens) matched_tokens = [] limit = len(keyword_parts) for idx in six.moves.range(limit): try: token = next(tokens) except StopIteration: return if (not token or token.name != "symbol" or token.value.lower() != keyword_parts[idx]): return matched_tokens.append(token) return TokenMatch(None, token.value, matched_tokens)
python
def multi_keyword(tokens, keyword_parts): """Match a case-insensitive keyword consisting of multiple tokens.""" tokens = iter(tokens) matched_tokens = [] limit = len(keyword_parts) for idx in six.moves.range(limit): try: token = next(tokens) except StopIteration: return if (not token or token.name != "symbol" or token.value.lower() != keyword_parts[idx]): return matched_tokens.append(token) return TokenMatch(None, token.value, matched_tokens)
[ "def", "multi_keyword", "(", "tokens", ",", "keyword_parts", ")", ":", "tokens", "=", "iter", "(", "tokens", ")", "matched_tokens", "=", "[", "]", "limit", "=", "len", "(", "keyword_parts", ")", "for", "idx", "in", "six", ".", "moves", ".", "range", "(", "limit", ")", ":", "try", ":", "token", "=", "next", "(", "tokens", ")", "except", "StopIteration", ":", "return", "if", "(", "not", "token", "or", "token", ".", "name", "!=", "\"symbol\"", "or", "token", ".", "value", ".", "lower", "(", ")", "!=", "keyword_parts", "[", "idx", "]", ")", ":", "return", "matched_tokens", ".", "append", "(", "token", ")", "return", "TokenMatch", "(", "None", ",", "token", ".", "value", ",", "matched_tokens", ")" ]
Match a case-insensitive keyword consisting of multiple tokens.
[ "Match", "a", "case", "-", "insensitive", "keyword", "consisting", "of", "multiple", "tokens", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L245-L263
google/dotty
efilter/parsers/common/grammar.py
prefix
def prefix(tokens, operator_table): """Match a prefix of an operator.""" operator, matched_tokens = operator_table.prefix.match(tokens) if operator: return TokenMatch(operator, None, matched_tokens)
python
def prefix(tokens, operator_table): """Match a prefix of an operator.""" operator, matched_tokens = operator_table.prefix.match(tokens) if operator: return TokenMatch(operator, None, matched_tokens)
[ "def", "prefix", "(", "tokens", ",", "operator_table", ")", ":", "operator", ",", "matched_tokens", "=", "operator_table", ".", "prefix", ".", "match", "(", "tokens", ")", "if", "operator", ":", "return", "TokenMatch", "(", "operator", ",", "None", ",", "matched_tokens", ")" ]
Match a prefix of an operator.
[ "Match", "a", "prefix", "of", "an", "operator", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L281-L285
google/dotty
efilter/parsers/common/grammar.py
infix
def infix(tokens, operator_table): """Match an infix of an operator.""" operator, matched_tokens = operator_table.infix.match(tokens) if operator: return TokenMatch(operator, None, matched_tokens)
python
def infix(tokens, operator_table): """Match an infix of an operator.""" operator, matched_tokens = operator_table.infix.match(tokens) if operator: return TokenMatch(operator, None, matched_tokens)
[ "def", "infix", "(", "tokens", ",", "operator_table", ")", ":", "operator", ",", "matched_tokens", "=", "operator_table", ".", "infix", ".", "match", "(", "tokens", ")", "if", "operator", ":", "return", "TokenMatch", "(", "operator", ",", "None", ",", "matched_tokens", ")" ]
Match an infix of an operator.
[ "Match", "an", "infix", "of", "an", "operator", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L288-L292
google/dotty
efilter/parsers/common/grammar.py
suffix
def suffix(tokens, operator_table): """Match a suffix of an operator.""" operator, matched_tokens = operator_table.suffix.match(tokens) if operator: return TokenMatch(operator, None, matched_tokens)
python
def suffix(tokens, operator_table): """Match a suffix of an operator.""" operator, matched_tokens = operator_table.suffix.match(tokens) if operator: return TokenMatch(operator, None, matched_tokens)
[ "def", "suffix", "(", "tokens", ",", "operator_table", ")", ":", "operator", ",", "matched_tokens", "=", "operator_table", ".", "suffix", ".", "match", "(", "tokens", ")", "if", "operator", ":", "return", "TokenMatch", "(", "operator", ",", "None", ",", "matched_tokens", ")" ]
Match a suffix of an operator.
[ "Match", "a", "suffix", "of", "an", "operator", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L295-L299
google/dotty
efilter/parsers/common/grammar.py
token_name
def token_name(tokens, expected): """Match a token name (type).""" try: token = next(iter(tokens)) except StopIteration: return if token and token.name == expected: return TokenMatch(None, token.value, (token,))
python
def token_name(tokens, expected): """Match a token name (type).""" try: token = next(iter(tokens)) except StopIteration: return if token and token.name == expected: return TokenMatch(None, token.value, (token,))
[ "def", "token_name", "(", "tokens", ",", "expected", ")", ":", "try", ":", "token", "=", "next", "(", "iter", "(", "tokens", ")", ")", "except", "StopIteration", ":", "return", "if", "token", "and", "token", ".", "name", "==", "expected", ":", "return", "TokenMatch", "(", "None", ",", "token", ".", "value", ",", "(", "token", ",", ")", ")" ]
Match a token name (type).
[ "Match", "a", "token", "name", "(", "type", ")", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L302-L310
google/dotty
efilter/parsers/common/grammar.py
match_tokens
def match_tokens(expected_tokens): """Generate a grammar function that will match 'expected_tokens' only.""" if isinstance(expected_tokens, Token): # Match a single token. def _grammar_func(tokens): try: next_token = next(iter(tokens)) except StopIteration: return if next_token == expected_tokens: return TokenMatch(None, next_token.value, (next_token,)) elif isinstance(expected_tokens, tuple): # Match multiple tokens. match_len = len(expected_tokens) def _grammar_func(tokens): upcoming = tuple(itertools.islice(tokens, match_len)) if upcoming == expected_tokens: return TokenMatch(None, None, upcoming) else: raise TypeError( "'expected_tokens' must be an instance of Token or a tuple " "thereof. Got %r." % expected_tokens) return _grammar_func
python
def match_tokens(expected_tokens): """Generate a grammar function that will match 'expected_tokens' only.""" if isinstance(expected_tokens, Token): # Match a single token. def _grammar_func(tokens): try: next_token = next(iter(tokens)) except StopIteration: return if next_token == expected_tokens: return TokenMatch(None, next_token.value, (next_token,)) elif isinstance(expected_tokens, tuple): # Match multiple tokens. match_len = len(expected_tokens) def _grammar_func(tokens): upcoming = tuple(itertools.islice(tokens, match_len)) if upcoming == expected_tokens: return TokenMatch(None, None, upcoming) else: raise TypeError( "'expected_tokens' must be an instance of Token or a tuple " "thereof. Got %r." % expected_tokens) return _grammar_func
[ "def", "match_tokens", "(", "expected_tokens", ")", ":", "if", "isinstance", "(", "expected_tokens", ",", "Token", ")", ":", "# Match a single token.", "def", "_grammar_func", "(", "tokens", ")", ":", "try", ":", "next_token", "=", "next", "(", "iter", "(", "tokens", ")", ")", "except", "StopIteration", ":", "return", "if", "next_token", "==", "expected_tokens", ":", "return", "TokenMatch", "(", "None", ",", "next_token", ".", "value", ",", "(", "next_token", ",", ")", ")", "elif", "isinstance", "(", "expected_tokens", ",", "tuple", ")", ":", "# Match multiple tokens.", "match_len", "=", "len", "(", "expected_tokens", ")", "def", "_grammar_func", "(", "tokens", ")", ":", "upcoming", "=", "tuple", "(", "itertools", ".", "islice", "(", "tokens", ",", "match_len", ")", ")", "if", "upcoming", "==", "expected_tokens", ":", "return", "TokenMatch", "(", "None", ",", "None", ",", "upcoming", ")", "else", ":", "raise", "TypeError", "(", "\"'expected_tokens' must be an instance of Token or a tuple \"", "\"thereof. Got %r.\"", "%", "expected_tokens", ")", "return", "_grammar_func" ]
Generate a grammar function that will match 'expected_tokens' only.
[ "Generate", "a", "grammar", "function", "that", "will", "match", "expected_tokens", "only", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/common/grammar.py#L313-L338
bradmontgomery/django-redis-metrics
redis_metrics/utils.py
set_metric
def set_metric(slug, value, category=None, expire=None, date=None): """Create/Increment a metric.""" get_r().set_metric(slug, value, category=category, expire=expire, date=date)
python
def set_metric(slug, value, category=None, expire=None, date=None): """Create/Increment a metric.""" get_r().set_metric(slug, value, category=category, expire=expire, date=date)
[ "def", "set_metric", "(", "slug", ",", "value", ",", "category", "=", "None", ",", "expire", "=", "None", ",", "date", "=", "None", ")", ":", "get_r", "(", ")", ".", "set_metric", "(", "slug", ",", "value", ",", "category", "=", "category", ",", "expire", "=", "expire", ",", "date", "=", "date", ")" ]
Create/Increment a metric.
[ "Create", "/", "Increment", "a", "metric", "." ]
train
https://github.com/bradmontgomery/django-redis-metrics/blob/2c92332920113d28c39234b949aa496b39a091d1/redis_metrics/utils.py#L18-L20
bradmontgomery/django-redis-metrics
redis_metrics/utils.py
metric
def metric(slug, num=1, category=None, expire=None, date=None): """Create/Increment a metric.""" get_r().metric(slug, num=num, category=category, expire=expire, date=date)
python
def metric(slug, num=1, category=None, expire=None, date=None): """Create/Increment a metric.""" get_r().metric(slug, num=num, category=category, expire=expire, date=date)
[ "def", "metric", "(", "slug", ",", "num", "=", "1", ",", "category", "=", "None", ",", "expire", "=", "None", ",", "date", "=", "None", ")", ":", "get_r", "(", ")", ".", "metric", "(", "slug", ",", "num", "=", "num", ",", "category", "=", "category", ",", "expire", "=", "expire", ",", "date", "=", "date", ")" ]
Create/Increment a metric.
[ "Create", "/", "Increment", "a", "metric", "." ]
train
https://github.com/bradmontgomery/django-redis-metrics/blob/2c92332920113d28c39234b949aa496b39a091d1/redis_metrics/utils.py#L23-L25
google/dotty
efilter/parsers/dottysql/parser.py
Parser.expression
def expression(self, previous_precedence=0): """An expression is an atom or an infix expression. Grammar (sort of, actually a precedence-climbing parser): expression = atom [ binary_operator expression ] . Args: previous_precedence: What operator precedence should we start with? """ lhs = self.atom() return self.operator(lhs, previous_precedence)
python
def expression(self, previous_precedence=0): """An expression is an atom or an infix expression. Grammar (sort of, actually a precedence-climbing parser): expression = atom [ binary_operator expression ] . Args: previous_precedence: What operator precedence should we start with? """ lhs = self.atom() return self.operator(lhs, previous_precedence)
[ "def", "expression", "(", "self", ",", "previous_precedence", "=", "0", ")", ":", "lhs", "=", "self", ".", "atom", "(", ")", "return", "self", ".", "operator", "(", "lhs", ",", "previous_precedence", ")" ]
An expression is an atom or an infix expression. Grammar (sort of, actually a precedence-climbing parser): expression = atom [ binary_operator expression ] . Args: previous_precedence: What operator precedence should we start with?
[ "An", "expression", "is", "an", "atom", "or", "an", "infix", "expression", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/dottysql/parser.py#L143-L154
google/dotty
efilter/parsers/dottysql/parser.py
Parser.atom
def atom(self): """Parse an atom, which is most things. Grammar: atom = [ prefix ] ( select_expression | any_expression | func_application | let_expr | var | literal | list | "(" expression ")" ) . """ # Parameter replacement with literals. if self.tokens.accept(grammar.param): return self.param() # Let expressions (let(x = 5, y = 10) x + y) if self.tokens.accept(grammar.let): return self.let() # At the top level, we try to see if we are recursing into an SQL query. if self.tokens.accept(grammar.select): return self.select() # A SELECT query can also start with 'ANY'. if self.tokens.accept(grammar.select_any): return self.select_any() # Explicitly reject any keywords from SQL other than SELECT and ANY. # If we don't do this they will match as valid symbols (variables) # and that might be confusing to the user. self.tokens.reject(grammar.sql_keyword) # Match if-else before other things that consume symbols. if self.tokens.accept(grammar.if_if): return self.if_if() # Operators must be matched first because the same symbols could also # be vars or applications. if self.tokens.accept(grammar.prefix): operator = self.tokens.matched.operator start = self.tokens.matched.start expr = self.expression(operator.precedence) return operator.handler(expr, start=start, end=expr.end, source=self.original) if self.tokens.accept(grammar.literal): return ast.Literal(self.tokens.matched.value, source=self.original, start=self.tokens.matched.start, end=self.tokens.matched.end) # Match builtin pseudo-functions before functions and vars to prevent # overrides. if self.tokens.accept(grammar.builtin): return self.builtin(self.tokens.matched.value) # Match applications before vars, because obviously. if self.tokens.accept(grammar.application): return self.application( ast.Var(self.tokens.matched.value, source=self.original, start=self.tokens.matched.start, end=self.tokens.matched.first.end)) if self.tokens.accept(common_grammar.symbol): return ast.Var(self.tokens.matched.value, source=self.original, start=self.tokens.matched.start, end=self.tokens.matched.end) if self.tokens.accept(common_grammar.lparen): # Parens will contain one or more expressions. If there are several # expressions, separated by commas, then they are a repeated value. # # Unlike lists, repeated values must all be of the same type, # otherwise evaluation of the query will fail at runtime (or # type-check time, for simple cases.) start = self.tokens.matched.start expressions = [self.expression()] while self.tokens.accept(common_grammar.comma): expressions.append(self.expression()) self.tokens.expect(common_grammar.rparen) if len(expressions) == 1: return expressions[0] else: return ast.Repeat(*expressions, source=self.original, start=start, end=self.tokens.matched.end) if self.tokens.accept(common_grammar.lbracket): return self.list() # We've run out of things we know the next atom could be. If there is # still input left then it's illegal syntax. If there is nothing then # the input cuts off when we still need an atom. Either is an error. if self.tokens.peek(0): return self.error( "Was not expecting %r here." % self.tokens.peek(0).name, start_token=self.tokens.peek(0)) else: return self.error("Unexpected end of input.")
python
def atom(self): """Parse an atom, which is most things. Grammar: atom = [ prefix ] ( select_expression | any_expression | func_application | let_expr | var | literal | list | "(" expression ")" ) . """ # Parameter replacement with literals. if self.tokens.accept(grammar.param): return self.param() # Let expressions (let(x = 5, y = 10) x + y) if self.tokens.accept(grammar.let): return self.let() # At the top level, we try to see if we are recursing into an SQL query. if self.tokens.accept(grammar.select): return self.select() # A SELECT query can also start with 'ANY'. if self.tokens.accept(grammar.select_any): return self.select_any() # Explicitly reject any keywords from SQL other than SELECT and ANY. # If we don't do this they will match as valid symbols (variables) # and that might be confusing to the user. self.tokens.reject(grammar.sql_keyword) # Match if-else before other things that consume symbols. if self.tokens.accept(grammar.if_if): return self.if_if() # Operators must be matched first because the same symbols could also # be vars or applications. if self.tokens.accept(grammar.prefix): operator = self.tokens.matched.operator start = self.tokens.matched.start expr = self.expression(operator.precedence) return operator.handler(expr, start=start, end=expr.end, source=self.original) if self.tokens.accept(grammar.literal): return ast.Literal(self.tokens.matched.value, source=self.original, start=self.tokens.matched.start, end=self.tokens.matched.end) # Match builtin pseudo-functions before functions and vars to prevent # overrides. if self.tokens.accept(grammar.builtin): return self.builtin(self.tokens.matched.value) # Match applications before vars, because obviously. if self.tokens.accept(grammar.application): return self.application( ast.Var(self.tokens.matched.value, source=self.original, start=self.tokens.matched.start, end=self.tokens.matched.first.end)) if self.tokens.accept(common_grammar.symbol): return ast.Var(self.tokens.matched.value, source=self.original, start=self.tokens.matched.start, end=self.tokens.matched.end) if self.tokens.accept(common_grammar.lparen): # Parens will contain one or more expressions. If there are several # expressions, separated by commas, then they are a repeated value. # # Unlike lists, repeated values must all be of the same type, # otherwise evaluation of the query will fail at runtime (or # type-check time, for simple cases.) start = self.tokens.matched.start expressions = [self.expression()] while self.tokens.accept(common_grammar.comma): expressions.append(self.expression()) self.tokens.expect(common_grammar.rparen) if len(expressions) == 1: return expressions[0] else: return ast.Repeat(*expressions, source=self.original, start=start, end=self.tokens.matched.end) if self.tokens.accept(common_grammar.lbracket): return self.list() # We've run out of things we know the next atom could be. If there is # still input left then it's illegal syntax. If there is nothing then # the input cuts off when we still need an atom. Either is an error. if self.tokens.peek(0): return self.error( "Was not expecting %r here." % self.tokens.peek(0).name, start_token=self.tokens.peek(0)) else: return self.error("Unexpected end of input.")
[ "def", "atom", "(", "self", ")", ":", "# Parameter replacement with literals.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "param", ")", ":", "return", "self", ".", "param", "(", ")", "# Let expressions (let(x = 5, y = 10) x + y)", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "let", ")", ":", "return", "self", ".", "let", "(", ")", "# At the top level, we try to see if we are recursing into an SQL query.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "select", ")", ":", "return", "self", ".", "select", "(", ")", "# A SELECT query can also start with 'ANY'.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "select_any", ")", ":", "return", "self", ".", "select_any", "(", ")", "# Explicitly reject any keywords from SQL other than SELECT and ANY.", "# If we don't do this they will match as valid symbols (variables)", "# and that might be confusing to the user.", "self", ".", "tokens", ".", "reject", "(", "grammar", ".", "sql_keyword", ")", "# Match if-else before other things that consume symbols.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "if_if", ")", ":", "return", "self", ".", "if_if", "(", ")", "# Operators must be matched first because the same symbols could also", "# be vars or applications.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "prefix", ")", ":", "operator", "=", "self", ".", "tokens", ".", "matched", ".", "operator", "start", "=", "self", ".", "tokens", ".", "matched", ".", "start", "expr", "=", "self", ".", "expression", "(", "operator", ".", "precedence", ")", "return", "operator", ".", "handler", "(", "expr", ",", "start", "=", "start", ",", "end", "=", "expr", ".", "end", ",", "source", "=", "self", ".", "original", ")", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "literal", ")", ":", "return", "ast", ".", "Literal", "(", "self", ".", "tokens", ".", "matched", ".", "value", ",", "source", "=", "self", ".", "original", ",", "start", "=", "self", ".", "tokens", ".", "matched", ".", "start", ",", "end", "=", "self", ".", "tokens", ".", "matched", ".", "end", ")", "# Match builtin pseudo-functions before functions and vars to prevent", "# overrides.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "builtin", ")", ":", "return", "self", ".", "builtin", "(", "self", ".", "tokens", ".", "matched", ".", "value", ")", "# Match applications before vars, because obviously.", "if", "self", ".", "tokens", ".", "accept", "(", "grammar", ".", "application", ")", ":", "return", "self", ".", "application", "(", "ast", ".", "Var", "(", "self", ".", "tokens", ".", "matched", ".", "value", ",", "source", "=", "self", ".", "original", ",", "start", "=", "self", ".", "tokens", ".", "matched", ".", "start", ",", "end", "=", "self", ".", "tokens", ".", "matched", ".", "first", ".", "end", ")", ")", "if", "self", ".", "tokens", ".", "accept", "(", "common_grammar", ".", "symbol", ")", ":", "return", "ast", ".", "Var", "(", "self", ".", "tokens", ".", "matched", ".", "value", ",", "source", "=", "self", ".", "original", ",", "start", "=", "self", ".", "tokens", ".", "matched", ".", "start", ",", "end", "=", "self", ".", "tokens", ".", "matched", ".", "end", ")", "if", "self", ".", "tokens", ".", "accept", "(", "common_grammar", ".", "lparen", ")", ":", "# Parens will contain one or more expressions. If there are several", "# expressions, separated by commas, then they are a repeated value.", "#", "# Unlike lists, repeated values must all be of the same type,", "# otherwise evaluation of the query will fail at runtime (or", "# type-check time, for simple cases.)", "start", "=", "self", ".", "tokens", ".", "matched", ".", "start", "expressions", "=", "[", "self", ".", "expression", "(", ")", "]", "while", "self", ".", "tokens", ".", "accept", "(", "common_grammar", ".", "comma", ")", ":", "expressions", ".", "append", "(", "self", ".", "expression", "(", ")", ")", "self", ".", "tokens", ".", "expect", "(", "common_grammar", ".", "rparen", ")", "if", "len", "(", "expressions", ")", "==", "1", ":", "return", "expressions", "[", "0", "]", "else", ":", "return", "ast", ".", "Repeat", "(", "*", "expressions", ",", "source", "=", "self", ".", "original", ",", "start", "=", "start", ",", "end", "=", "self", ".", "tokens", ".", "matched", ".", "end", ")", "if", "self", ".", "tokens", ".", "accept", "(", "common_grammar", ".", "lbracket", ")", ":", "return", "self", ".", "list", "(", ")", "# We've run out of things we know the next atom could be. If there is", "# still input left then it's illegal syntax. If there is nothing then", "# the input cuts off when we still need an atom. Either is an error.", "if", "self", ".", "tokens", ".", "peek", "(", "0", ")", ":", "return", "self", ".", "error", "(", "\"Was not expecting %r here.\"", "%", "self", ".", "tokens", ".", "peek", "(", "0", ")", ".", "name", ",", "start_token", "=", "self", ".", "tokens", ".", "peek", "(", "0", ")", ")", "else", ":", "return", "self", ".", "error", "(", "\"Unexpected end of input.\"", ")" ]
Parse an atom, which is most things. Grammar: atom = [ prefix ] ( select_expression | any_expression | func_application | let_expr | var | literal | list | "(" expression ")" ) .
[ "Parse", "an", "atom", "which", "is", "most", "things", "." ]
train
https://github.com/google/dotty/blob/b145131499be0c4b755fc2e2ac19be11a50bce6a/efilter/parsers/dottysql/parser.py#L156-L259