partition
stringclasses
3 values
func_name
stringlengths
1
134
docstring
stringlengths
1
46.9k
path
stringlengths
4
223
original_string
stringlengths
75
104k
code
stringlengths
75
104k
docstring_tokens
listlengths
1
1.97k
repo
stringlengths
7
55
language
stringclasses
1 value
url
stringlengths
87
315
code_tokens
listlengths
19
28.4k
sha
stringlengths
40
40
valid
Packager.package_for_editor_signavio
Adds the SVG files to the archive for this BPMN file.
SpiffWorkflow/bpmn/serializer/Packager.py
def package_for_editor_signavio(self, spec, filename): """ Adds the SVG files to the archive for this BPMN file. """ signavio_file = filename[:-len('.bpmn20.xml')] + '.signavio.xml' if os.path.exists(signavio_file): self.write_file_to_package_zip( "src/" + self._get_zip_path(signavio_file), signavio_file) f = open(signavio_file, 'r') try: signavio_tree = ET.parse(f) finally: f.close() svg_node = one(signavio_tree.findall('.//svg-representation')) self.write_to_package_zip("%s.svg" % spec.name, svg_node.text)
def package_for_editor_signavio(self, spec, filename): """ Adds the SVG files to the archive for this BPMN file. """ signavio_file = filename[:-len('.bpmn20.xml')] + '.signavio.xml' if os.path.exists(signavio_file): self.write_file_to_package_zip( "src/" + self._get_zip_path(signavio_file), signavio_file) f = open(signavio_file, 'r') try: signavio_tree = ET.parse(f) finally: f.close() svg_node = one(signavio_tree.findall('.//svg-representation')) self.write_to_package_zip("%s.svg" % spec.name, svg_node.text)
[ "Adds", "the", "SVG", "files", "to", "the", "archive", "for", "this", "BPMN", "file", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L322-L337
[ "def", "package_for_editor_signavio", "(", "self", ",", "spec", ",", "filename", ")", ":", "signavio_file", "=", "filename", "[", ":", "-", "len", "(", "'.bpmn20.xml'", ")", "]", "+", "'.signavio.xml'", "if", "os", ".", "path", ".", "exists", "(", "signavio_file", ")", ":", "self", ".", "write_file_to_package_zip", "(", "\"src/\"", "+", "self", ".", "_get_zip_path", "(", "signavio_file", ")", ",", "signavio_file", ")", "f", "=", "open", "(", "signavio_file", ",", "'r'", ")", "try", ":", "signavio_tree", "=", "ET", ".", "parse", "(", "f", ")", "finally", ":", "f", ".", "close", "(", ")", "svg_node", "=", "one", "(", "signavio_tree", ".", "findall", "(", "'.//svg-representation'", ")", ")", "self", ".", "write_to_package_zip", "(", "\"%s.svg\"", "%", "spec", ".", "name", ",", "svg_node", ".", "text", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.write_meta_data
Writes the metadata.ini file to the archive.
SpiffWorkflow/bpmn/serializer/Packager.py
def write_meta_data(self): """ Writes the metadata.ini file to the archive. """ config = configparser.ConfigParser() config.add_section('MetaData') config.set('MetaData', 'entry_point_process', self.wf_spec.name) if self.editor: config.set('MetaData', 'editor', self.editor) for k, v in self.meta_data: config.set('MetaData', k, v) if not self.PARSER_CLASS == BpmnParser: config.set('MetaData', 'parser_class_module', inspect.getmodule(self.PARSER_CLASS).__name__) config.set('MetaData', 'parser_class', self.PARSER_CLASS.__name__) ini = StringIO() config.write(ini) self.write_to_package_zip(self.METADATA_FILE, ini.getvalue())
def write_meta_data(self): """ Writes the metadata.ini file to the archive. """ config = configparser.ConfigParser() config.add_section('MetaData') config.set('MetaData', 'entry_point_process', self.wf_spec.name) if self.editor: config.set('MetaData', 'editor', self.editor) for k, v in self.meta_data: config.set('MetaData', k, v) if not self.PARSER_CLASS == BpmnParser: config.set('MetaData', 'parser_class_module', inspect.getmodule(self.PARSER_CLASS).__name__) config.set('MetaData', 'parser_class', self.PARSER_CLASS.__name__) ini = StringIO() config.write(ini) self.write_to_package_zip(self.METADATA_FILE, ini.getvalue())
[ "Writes", "the", "metadata", ".", "ini", "file", "to", "the", "archive", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L339-L360
[ "def", "write_meta_data", "(", "self", ")", ":", "config", "=", "configparser", ".", "ConfigParser", "(", ")", "config", ".", "add_section", "(", "'MetaData'", ")", "config", ".", "set", "(", "'MetaData'", ",", "'entry_point_process'", ",", "self", ".", "wf_spec", ".", "name", ")", "if", "self", ".", "editor", ":", "config", ".", "set", "(", "'MetaData'", ",", "'editor'", ",", "self", ".", "editor", ")", "for", "k", ",", "v", "in", "self", ".", "meta_data", ":", "config", ".", "set", "(", "'MetaData'", ",", "k", ",", "v", ")", "if", "not", "self", ".", "PARSER_CLASS", "==", "BpmnParser", ":", "config", ".", "set", "(", "'MetaData'", ",", "'parser_class_module'", ",", "inspect", ".", "getmodule", "(", "self", ".", "PARSER_CLASS", ")", ".", "__name__", ")", "config", ".", "set", "(", "'MetaData'", ",", "'parser_class'", ",", "self", ".", "PARSER_CLASS", ".", "__name__", ")", "ini", "=", "StringIO", "(", ")", "config", ".", "write", "(", "ini", ")", "self", ".", "write_to_package_zip", "(", "self", ".", "METADATA_FILE", ",", "ini", ".", "getvalue", "(", ")", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.add_main_options
Override in subclass if required.
SpiffWorkflow/bpmn/serializer/Packager.py
def add_main_options(cls, parser): """ Override in subclass if required. """ parser.add_option("-o", "--output", dest="package_file", help="create the BPMN package in the specified file") parser.add_option("-p", "--process", dest="entry_point_process", help="specify the entry point process") parser.add_option("-c", "--config-file", dest="config_file", help="specify a config file to use") parser.add_option( "-i", "--initialise-config-file", action="store_true", dest="init_config_file", default=False, help="create a new config file from the specified options") group = OptionGroup(parser, "BPMN Editor Options", "These options are not required, but may be " " provided to activate special features of " "supported BPMN editors.") group.add_option("--editor", dest="editor", help="editors with special support: signavio") parser.add_option_group(group)
def add_main_options(cls, parser): """ Override in subclass if required. """ parser.add_option("-o", "--output", dest="package_file", help="create the BPMN package in the specified file") parser.add_option("-p", "--process", dest="entry_point_process", help="specify the entry point process") parser.add_option("-c", "--config-file", dest="config_file", help="specify a config file to use") parser.add_option( "-i", "--initialise-config-file", action="store_true", dest="init_config_file", default=False, help="create a new config file from the specified options") group = OptionGroup(parser, "BPMN Editor Options", "These options are not required, but may be " " provided to activate special features of " "supported BPMN editors.") group.add_option("--editor", dest="editor", help="editors with special support: signavio") parser.add_option_group(group)
[ "Override", "in", "subclass", "if", "required", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L389-L410
[ "def", "add_main_options", "(", "cls", ",", "parser", ")", ":", "parser", ".", "add_option", "(", "\"-o\"", ",", "\"--output\"", ",", "dest", "=", "\"package_file\"", ",", "help", "=", "\"create the BPMN package in the specified file\"", ")", "parser", ".", "add_option", "(", "\"-p\"", ",", "\"--process\"", ",", "dest", "=", "\"entry_point_process\"", ",", "help", "=", "\"specify the entry point process\"", ")", "parser", ".", "add_option", "(", "\"-c\"", ",", "\"--config-file\"", ",", "dest", "=", "\"config_file\"", ",", "help", "=", "\"specify a config file to use\"", ")", "parser", ".", "add_option", "(", "\"-i\"", ",", "\"--initialise-config-file\"", ",", "action", "=", "\"store_true\"", ",", "dest", "=", "\"init_config_file\"", ",", "default", "=", "False", ",", "help", "=", "\"create a new config file from the specified options\"", ")", "group", "=", "OptionGroup", "(", "parser", ",", "\"BPMN Editor Options\"", ",", "\"These options are not required, but may be \"", "\" provided to activate special features of \"", "\"supported BPMN editors.\"", ")", "group", ".", "add_option", "(", "\"--editor\"", ",", "dest", "=", "\"editor\"", ",", "help", "=", "\"editors with special support: signavio\"", ")", "parser", ".", "add_option_group", "(", "group", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.add_additional_options
Override in subclass if required.
SpiffWorkflow/bpmn/serializer/Packager.py
def add_additional_options(cls, parser): """ Override in subclass if required. """ group = OptionGroup(parser, "Target Engine Options", "These options are not required, but may be " "provided if a specific " "BPMN application engine is targeted.") group.add_option("-e", "--target-engine", dest="target_engine", help="target the specified BPMN application engine") group.add_option( "-t", "--target-version", dest="target_engine_version", help="target the specified version of the BPMN application engine") parser.add_option_group(group)
def add_additional_options(cls, parser): """ Override in subclass if required. """ group = OptionGroup(parser, "Target Engine Options", "These options are not required, but may be " "provided if a specific " "BPMN application engine is targeted.") group.add_option("-e", "--target-engine", dest="target_engine", help="target the specified BPMN application engine") group.add_option( "-t", "--target-version", dest="target_engine_version", help="target the specified version of the BPMN application engine") parser.add_option_group(group)
[ "Override", "in", "subclass", "if", "required", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L413-L426
[ "def", "add_additional_options", "(", "cls", ",", "parser", ")", ":", "group", "=", "OptionGroup", "(", "parser", ",", "\"Target Engine Options\"", ",", "\"These options are not required, but may be \"", "\"provided if a specific \"", "\"BPMN application engine is targeted.\"", ")", "group", ".", "add_option", "(", "\"-e\"", ",", "\"--target-engine\"", ",", "dest", "=", "\"target_engine\"", ",", "help", "=", "\"target the specified BPMN application engine\"", ")", "group", ".", "add_option", "(", "\"-t\"", ",", "\"--target-version\"", ",", "dest", "=", "\"target_engine_version\"", ",", "help", "=", "\"target the specified version of the BPMN application engine\"", ")", "parser", ".", "add_option_group", "(", "group", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.check_args
Override in subclass if required.
SpiffWorkflow/bpmn/serializer/Packager.py
def check_args(cls, config, options, args, parser, package_file=None): """ Override in subclass if required. """ if not args: parser.error("no input files specified") if not (package_file or options.package_file): parser.error("no package file specified") if not options.entry_point_process: parser.error("no entry point process specified")
def check_args(cls, config, options, args, parser, package_file=None): """ Override in subclass if required. """ if not args: parser.error("no input files specified") if not (package_file or options.package_file): parser.error("no package file specified") if not options.entry_point_process: parser.error("no entry point process specified")
[ "Override", "in", "subclass", "if", "required", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L429-L438
[ "def", "check_args", "(", "cls", ",", "config", ",", "options", ",", "args", ",", "parser", ",", "package_file", "=", "None", ")", ":", "if", "not", "args", ":", "parser", ".", "error", "(", "\"no input files specified\"", ")", "if", "not", "(", "package_file", "or", "options", ".", "package_file", ")", ":", "parser", ".", "error", "(", "\"no package file specified\"", ")", "if", "not", "options", ".", "entry_point_process", ":", "parser", ".", "error", "(", "\"no entry point process specified\"", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.merge_options_and_config
Override in subclass if required.
SpiffWorkflow/bpmn/serializer/Packager.py
def merge_options_and_config(cls, config, options, args): """ Override in subclass if required. """ if args: config.set(CONFIG_SECTION_NAME, 'input_files', ','.join(args)) elif config.has_option(CONFIG_SECTION_NAME, 'input_files'): for i in config.get(CONFIG_SECTION_NAME, 'input_files').split(','): if not os.path.isabs(i): i = os.path.abspath( os.path.join(os.path.dirname(options.config_file), i)) args.append(i) cls.merge_option_and_config_str('package_file', config, options) cls.merge_option_and_config_str('entry_point_process', config, options) cls.merge_option_and_config_str('target_engine', config, options) cls.merge_option_and_config_str( 'target_engine_version', config, options) cls.merge_option_and_config_str('editor', config, options)
def merge_options_and_config(cls, config, options, args): """ Override in subclass if required. """ if args: config.set(CONFIG_SECTION_NAME, 'input_files', ','.join(args)) elif config.has_option(CONFIG_SECTION_NAME, 'input_files'): for i in config.get(CONFIG_SECTION_NAME, 'input_files').split(','): if not os.path.isabs(i): i = os.path.abspath( os.path.join(os.path.dirname(options.config_file), i)) args.append(i) cls.merge_option_and_config_str('package_file', config, options) cls.merge_option_and_config_str('entry_point_process', config, options) cls.merge_option_and_config_str('target_engine', config, options) cls.merge_option_and_config_str( 'target_engine_version', config, options) cls.merge_option_and_config_str('editor', config, options)
[ "Override", "in", "subclass", "if", "required", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L441-L459
[ "def", "merge_options_and_config", "(", "cls", ",", "config", ",", "options", ",", "args", ")", ":", "if", "args", ":", "config", ".", "set", "(", "CONFIG_SECTION_NAME", ",", "'input_files'", ",", "','", ".", "join", "(", "args", ")", ")", "elif", "config", ".", "has_option", "(", "CONFIG_SECTION_NAME", ",", "'input_files'", ")", ":", "for", "i", "in", "config", ".", "get", "(", "CONFIG_SECTION_NAME", ",", "'input_files'", ")", ".", "split", "(", "','", ")", ":", "if", "not", "os", ".", "path", ".", "isabs", "(", "i", ")", ":", "i", "=", "os", ".", "path", ".", "abspath", "(", "os", ".", "path", ".", "join", "(", "os", ".", "path", ".", "dirname", "(", "options", ".", "config_file", ")", ",", "i", ")", ")", "args", ".", "append", "(", "i", ")", "cls", ".", "merge_option_and_config_str", "(", "'package_file'", ",", "config", ",", "options", ")", "cls", ".", "merge_option_and_config_str", "(", "'entry_point_process'", ",", "config", ",", "options", ")", "cls", ".", "merge_option_and_config_str", "(", "'target_engine'", ",", "config", ",", "options", ")", "cls", ".", "merge_option_and_config_str", "(", "'target_engine_version'", ",", "config", ",", "options", ")", "cls", ".", "merge_option_and_config_str", "(", "'editor'", ",", "config", ",", "options", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.merge_option_and_config_str
Utility method to merge an option and config, with the option taking " precedence
SpiffWorkflow/bpmn/serializer/Packager.py
def merge_option_and_config_str(cls, option_name, config, options): """ Utility method to merge an option and config, with the option taking " precedence """ opt = getattr(options, option_name, None) if opt: config.set(CONFIG_SECTION_NAME, option_name, opt) elif config.has_option(CONFIG_SECTION_NAME, option_name): setattr(options, option_name, config.get( CONFIG_SECTION_NAME, option_name))
def merge_option_and_config_str(cls, option_name, config, options): """ Utility method to merge an option and config, with the option taking " precedence """ opt = getattr(options, option_name, None) if opt: config.set(CONFIG_SECTION_NAME, option_name, opt) elif config.has_option(CONFIG_SECTION_NAME, option_name): setattr(options, option_name, config.get( CONFIG_SECTION_NAME, option_name))
[ "Utility", "method", "to", "merge", "an", "option", "and", "config", "with", "the", "option", "taking", "precedence" ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L462-L473
[ "def", "merge_option_and_config_str", "(", "cls", ",", "option_name", ",", "config", ",", "options", ")", ":", "opt", "=", "getattr", "(", "options", ",", "option_name", ",", "None", ")", "if", "opt", ":", "config", ".", "set", "(", "CONFIG_SECTION_NAME", ",", "option_name", ",", "opt", ")", "elif", "config", ".", "has_option", "(", "CONFIG_SECTION_NAME", ",", "option_name", ")", ":", "setattr", "(", "options", ",", "option_name", ",", "config", ".", "get", "(", "CONFIG_SECTION_NAME", ",", "option_name", ")", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Packager.create_meta_data
Override in subclass if required.
SpiffWorkflow/bpmn/serializer/Packager.py
def create_meta_data(cls, options, args, parser): """ Override in subclass if required. """ meta_data = [] meta_data.append(('spiff_version', cls.get_version())) if options.target_engine: meta_data.append(('target_engine', options.target_engine)) if options.target_engine: meta_data.append( ('target_engine_version', options.target_engine_version)) return meta_data
def create_meta_data(cls, options, args, parser): """ Override in subclass if required. """ meta_data = [] meta_data.append(('spiff_version', cls.get_version())) if options.target_engine: meta_data.append(('target_engine', options.target_engine)) if options.target_engine: meta_data.append( ('target_engine_version', options.target_engine_version)) return meta_data
[ "Override", "in", "subclass", "if", "required", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/serializer/Packager.py#L476-L487
[ "def", "create_meta_data", "(", "cls", ",", "options", ",", "args", ",", "parser", ")", ":", "meta_data", "=", "[", "]", "meta_data", ".", "append", "(", "(", "'spiff_version'", ",", "cls", ".", "get_version", "(", ")", ")", ")", "if", "options", ".", "target_engine", ":", "meta_data", ".", "append", "(", "(", "'target_engine'", ",", "options", ".", "target_engine", ")", ")", "if", "options", ".", "target_engine", ":", "meta_data", ".", "append", "(", "(", "'target_engine_version'", ",", "options", ".", "target_engine_version", ")", ")", "return", "meta_data" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
ProcessParser.parse_node
Parses the specified child task node, and returns the task spec. This can be called by a TaskParser instance, that is owned by this ProcessParser.
SpiffWorkflow/bpmn/parser/ProcessParser.py
def parse_node(self, node): """ Parses the specified child task node, and returns the task spec. This can be called by a TaskParser instance, that is owned by this ProcessParser. """ if node.get('id') in self.parsed_nodes: return self.parsed_nodes[node.get('id')] (node_parser, spec_class) = self.parser._get_parser_class(node.tag) if not node_parser or not spec_class: raise ValidationException( "There is no support implemented for this task type.", node=node, filename=self.filename) np = node_parser(self, spec_class, node) task_spec = np.parse_node() return task_spec
def parse_node(self, node): """ Parses the specified child task node, and returns the task spec. This can be called by a TaskParser instance, that is owned by this ProcessParser. """ if node.get('id') in self.parsed_nodes: return self.parsed_nodes[node.get('id')] (node_parser, spec_class) = self.parser._get_parser_class(node.tag) if not node_parser or not spec_class: raise ValidationException( "There is no support implemented for this task type.", node=node, filename=self.filename) np = node_parser(self, spec_class, node) task_spec = np.parse_node() return task_spec
[ "Parses", "the", "specified", "child", "task", "node", "and", "returns", "the", "task", "spec", ".", "This", "can", "be", "called", "by", "a", "TaskParser", "instance", "that", "is", "owned", "by", "this", "ProcessParser", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/parser/ProcessParser.py#L69-L87
[ "def", "parse_node", "(", "self", ",", "node", ")", ":", "if", "node", ".", "get", "(", "'id'", ")", "in", "self", ".", "parsed_nodes", ":", "return", "self", ".", "parsed_nodes", "[", "node", ".", "get", "(", "'id'", ")", "]", "(", "node_parser", ",", "spec_class", ")", "=", "self", ".", "parser", ".", "_get_parser_class", "(", "node", ".", "tag", ")", "if", "not", "node_parser", "or", "not", "spec_class", ":", "raise", "ValidationException", "(", "\"There is no support implemented for this task type.\"", ",", "node", "=", "node", ",", "filename", "=", "self", ".", "filename", ")", "np", "=", "node_parser", "(", "self", ",", "spec_class", ",", "node", ")", "task_spec", "=", "np", ".", "parse_node", "(", ")", "return", "task_spec" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
ProcessParser.get_spec
Parse this process (if it has not already been parsed), and return the workflow spec.
SpiffWorkflow/bpmn/parser/ProcessParser.py
def get_spec(self): """ Parse this process (if it has not already been parsed), and return the workflow spec. """ if self.is_parsed: return self.spec if self.parsing_started: raise NotImplementedError( 'Recursive call Activities are not supported.') self._parse() return self.get_spec()
def get_spec(self): """ Parse this process (if it has not already been parsed), and return the workflow spec. """ if self.is_parsed: return self.spec if self.parsing_started: raise NotImplementedError( 'Recursive call Activities are not supported.') self._parse() return self.get_spec()
[ "Parse", "this", "process", "(", "if", "it", "has", "not", "already", "been", "parsed", ")", "and", "return", "the", "workflow", "spec", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/parser/ProcessParser.py#L118-L129
[ "def", "get_spec", "(", "self", ")", ":", "if", "self", ".", "is_parsed", ":", "return", "self", ".", "spec", "if", "self", ".", "parsing_started", ":", "raise", "NotImplementedError", "(", "'Recursive call Activities are not supported.'", ")", "self", ".", "_parse", "(", ")", "return", "self", ".", "get_spec", "(", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
MultiInstance._on_trigger
May be called after execute() was already completed to create an additional outbound task.
SpiffWorkflow/specs/MultiInstance.py
def _on_trigger(self, task_spec): """ May be called after execute() was already completed to create an additional outbound task. """ # Find a Task for this TaskSpec. my_task = self._find_my_task(task_spec) if my_task._has_state(Task.COMPLETED): state = Task.READY else: state = Task.FUTURE for output in self.outputs: new_task = my_task._add_child(output, state) new_task.triggered = True output._predict(new_task)
def _on_trigger(self, task_spec): """ May be called after execute() was already completed to create an additional outbound task. """ # Find a Task for this TaskSpec. my_task = self._find_my_task(task_spec) if my_task._has_state(Task.COMPLETED): state = Task.READY else: state = Task.FUTURE for output in self.outputs: new_task = my_task._add_child(output, state) new_task.triggered = True output._predict(new_task)
[ "May", "be", "called", "after", "execute", "()", "was", "already", "completed", "to", "create", "an", "additional", "outbound", "task", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/MultiInstance.py#L63-L77
[ "def", "_on_trigger", "(", "self", ",", "task_spec", ")", ":", "# Find a Task for this TaskSpec.", "my_task", "=", "self", ".", "_find_my_task", "(", "task_spec", ")", "if", "my_task", ".", "_has_state", "(", "Task", ".", "COMPLETED", ")", ":", "state", "=", "Task", ".", "READY", "else", ":", "state", "=", "Task", ".", "FUTURE", "for", "output", "in", "self", ".", "outputs", ":", "new_task", "=", "my_task", ".", "_add_child", "(", "output", ",", "state", ")", "new_task", ".", "triggered", "=", "True", "output", ".", "_predict", "(", "new_task", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
ThreadSplit.connect
Connect the *following* task to this one. In other words, the given task is added as an output task. task -- the task to connect to.
SpiffWorkflow/specs/ThreadSplit.py
def connect(self, task_spec): """ Connect the *following* task to this one. In other words, the given task is added as an output task. task -- the task to connect to. """ self.thread_starter.outputs.append(task_spec) task_spec._connect_notify(self.thread_starter)
def connect(self, task_spec): """ Connect the *following* task to this one. In other words, the given task is added as an output task. task -- the task to connect to. """ self.thread_starter.outputs.append(task_spec) task_spec._connect_notify(self.thread_starter)
[ "Connect", "the", "*", "following", "*", "task", "to", "this", "one", ".", "In", "other", "words", "the", "given", "task", "is", "added", "as", "an", "output", "task", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/ThreadSplit.py#L70-L78
[ "def", "connect", "(", "self", ",", "task_spec", ")", ":", "self", ".", "thread_starter", ".", "outputs", ".", "append", "(", "task_spec", ")", "task_spec", ".", "_connect_notify", "(", "self", ".", "thread_starter", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
ThreadSplit._get_activated_tasks
Returns the list of tasks that were activated in the previous call of execute(). Only returns tasks that point towards the destination task, i.e. those which have destination as a descendant. my_task -- the task of this TaskSpec destination -- the child task
SpiffWorkflow/specs/ThreadSplit.py
def _get_activated_tasks(self, my_task, destination): """ Returns the list of tasks that were activated in the previous call of execute(). Only returns tasks that point towards the destination task, i.e. those which have destination as a descendant. my_task -- the task of this TaskSpec destination -- the child task """ task = destination._find_ancestor(self.thread_starter) return self.thread_starter._get_activated_tasks(task, destination)
def _get_activated_tasks(self, my_task, destination): """ Returns the list of tasks that were activated in the previous call of execute(). Only returns tasks that point towards the destination task, i.e. those which have destination as a descendant. my_task -- the task of this TaskSpec destination -- the child task """ task = destination._find_ancestor(self.thread_starter) return self.thread_starter._get_activated_tasks(task, destination)
[ "Returns", "the", "list", "of", "tasks", "that", "were", "activated", "in", "the", "previous", "call", "of", "execute", "()", ".", "Only", "returns", "tasks", "that", "point", "towards", "the", "destination", "task", "i", ".", "e", ".", "those", "which", "have", "destination", "as", "a", "descendant", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/ThreadSplit.py#L80-L91
[ "def", "_get_activated_tasks", "(", "self", ",", "my_task", ",", "destination", ")", ":", "task", "=", "destination", ".", "_find_ancestor", "(", "self", ".", "thread_starter", ")", "return", "self", ".", "thread_starter", ".", "_get_activated_tasks", "(", "task", ",", "destination", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
ThreadSplit._on_trigger
May be called after execute() was already completed to create an additional outbound task.
SpiffWorkflow/specs/ThreadSplit.py
def _on_trigger(self, my_task): """ May be called after execute() was already completed to create an additional outbound task. """ for output in self.outputs: new_task = my_task.add_child(output, Task.READY) new_task.triggered = True
def _on_trigger(self, my_task): """ May be called after execute() was already completed to create an additional outbound task. """ for output in self.outputs: new_task = my_task.add_child(output, Task.READY) new_task.triggered = True
[ "May", "be", "called", "after", "execute", "()", "was", "already", "completed", "to", "create", "an", "additional", "outbound", "task", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/ThreadSplit.py#L102-L109
[ "def", "_on_trigger", "(", "self", ",", "my_task", ")", ":", "for", "output", "in", "self", ".", "outputs", ":", "new_task", "=", "my_task", ".", "add_child", "(", "output", ",", "Task", ".", "READY", ")", "new_task", ".", "triggered", "=", "True" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_assign
Reads the "pre-assign" or "post-assign" tag from the given node. start_node -- the xml node (xml.dom.minidom.Node)
SpiffWorkflow/serializer/prettyxml.py
def deserialize_assign(self, workflow, start_node): """ Reads the "pre-assign" or "post-assign" tag from the given node. start_node -- the xml node (xml.dom.minidom.Node) """ name = start_node.getAttribute('name') attrib = start_node.getAttribute('field') value = start_node.getAttribute('value') kwargs = {} if name == '': _exc('name attribute required') if attrib != '' and value != '': _exc('Both, field and right-value attributes found') elif attrib == '' and value == '': _exc('field or value attribute required') elif value != '': kwargs['right'] = value else: kwargs['right_attribute'] = attrib return operators.Assign(name, **kwargs)
def deserialize_assign(self, workflow, start_node): """ Reads the "pre-assign" or "post-assign" tag from the given node. start_node -- the xml node (xml.dom.minidom.Node) """ name = start_node.getAttribute('name') attrib = start_node.getAttribute('field') value = start_node.getAttribute('value') kwargs = {} if name == '': _exc('name attribute required') if attrib != '' and value != '': _exc('Both, field and right-value attributes found') elif attrib == '' and value == '': _exc('field or value attribute required') elif value != '': kwargs['right'] = value else: kwargs['right_attribute'] = attrib return operators.Assign(name, **kwargs)
[ "Reads", "the", "pre", "-", "assign", "or", "post", "-", "assign", "tag", "from", "the", "given", "node", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L50-L70
[ "def", "deserialize_assign", "(", "self", ",", "workflow", ",", "start_node", ")", ":", "name", "=", "start_node", ".", "getAttribute", "(", "'name'", ")", "attrib", "=", "start_node", ".", "getAttribute", "(", "'field'", ")", "value", "=", "start_node", ".", "getAttribute", "(", "'value'", ")", "kwargs", "=", "{", "}", "if", "name", "==", "''", ":", "_exc", "(", "'name attribute required'", ")", "if", "attrib", "!=", "''", "and", "value", "!=", "''", ":", "_exc", "(", "'Both, field and right-value attributes found'", ")", "elif", "attrib", "==", "''", "and", "value", "==", "''", ":", "_exc", "(", "'field or value attribute required'", ")", "elif", "value", "!=", "''", ":", "kwargs", "[", "'right'", "]", "=", "value", "else", ":", "kwargs", "[", "'right_attribute'", "]", "=", "attrib", "return", "operators", ".", "Assign", "(", "name", ",", "*", "*", "kwargs", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_data
Reads a "data" or "define" tag from the given node. start_node -- the xml node (xml.dom.minidom.Node)
SpiffWorkflow/serializer/prettyxml.py
def deserialize_data(self, workflow, start_node): """ Reads a "data" or "define" tag from the given node. start_node -- the xml node (xml.dom.minidom.Node) """ name = start_node.getAttribute('name') value = start_node.getAttribute('value') return name, value
def deserialize_data(self, workflow, start_node): """ Reads a "data" or "define" tag from the given node. start_node -- the xml node (xml.dom.minidom.Node) """ name = start_node.getAttribute('name') value = start_node.getAttribute('value') return name, value
[ "Reads", "a", "data", "or", "define", "tag", "from", "the", "given", "node", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L72-L80
[ "def", "deserialize_data", "(", "self", ",", "workflow", ",", "start_node", ")", ":", "name", "=", "start_node", ".", "getAttribute", "(", "'name'", ")", "value", "=", "start_node", ".", "getAttribute", "(", "'value'", ")", "return", "name", ",", "value" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_assign_list
Reads a list of assignments from the given node. workflow -- the workflow start_node -- the xml structure (xml.dom.minidom.Node)
SpiffWorkflow/serializer/prettyxml.py
def deserialize_assign_list(self, workflow, start_node): """ Reads a list of assignments from the given node. workflow -- the workflow start_node -- the xml structure (xml.dom.minidom.Node) """ # Collect all information. assignments = [] for node in start_node.childNodes: if node.nodeType != minidom.Node.ELEMENT_NODE: continue if node.nodeName.lower() == 'assign': assignments.append(self.deserialize_assign(workflow, node)) else: _exc('Unknown node: %s' % node.nodeName) return assignments
def deserialize_assign_list(self, workflow, start_node): """ Reads a list of assignments from the given node. workflow -- the workflow start_node -- the xml structure (xml.dom.minidom.Node) """ # Collect all information. assignments = [] for node in start_node.childNodes: if node.nodeType != minidom.Node.ELEMENT_NODE: continue if node.nodeName.lower() == 'assign': assignments.append(self.deserialize_assign(workflow, node)) else: _exc('Unknown node: %s' % node.nodeName) return assignments
[ "Reads", "a", "list", "of", "assignments", "from", "the", "given", "node", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L82-L98
[ "def", "deserialize_assign_list", "(", "self", ",", "workflow", ",", "start_node", ")", ":", "# Collect all information.", "assignments", "=", "[", "]", "for", "node", "in", "start_node", ".", "childNodes", ":", "if", "node", ".", "nodeType", "!=", "minidom", ".", "Node", ".", "ELEMENT_NODE", ":", "continue", "if", "node", ".", "nodeName", ".", "lower", "(", ")", "==", "'assign'", ":", "assignments", ".", "append", "(", "self", ".", "deserialize_assign", "(", "workflow", ",", "node", ")", ")", "else", ":", "_exc", "(", "'Unknown node: %s'", "%", "node", ".", "nodeName", ")", "return", "assignments" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_logical
Reads the logical tag from the given node, returns a Condition object. node -- the xml node (xml.dom.minidom.Node)
SpiffWorkflow/serializer/prettyxml.py
def deserialize_logical(self, node): """ Reads the logical tag from the given node, returns a Condition object. node -- the xml node (xml.dom.minidom.Node) """ term1_attrib = node.getAttribute('left-field') term1_value = node.getAttribute('left-value') op = node.nodeName.lower() term2_attrib = node.getAttribute('right-field') term2_value = node.getAttribute('right-value') if op not in _op_map: _exc('Invalid operator') if term1_attrib != '' and term1_value != '': _exc('Both, left-field and left-value attributes found') elif term1_attrib == '' and term1_value == '': _exc('left-field or left-value attribute required') elif term1_value != '': left = term1_value else: left = operators.Attrib(term1_attrib) if term2_attrib != '' and term2_value != '': _exc('Both, right-field and right-value attributes found') elif term2_attrib == '' and term2_value == '': _exc('right-field or right-value attribute required') elif term2_value != '': right = term2_value else: right = operators.Attrib(term2_attrib) return _op_map[op](left, right)
def deserialize_logical(self, node): """ Reads the logical tag from the given node, returns a Condition object. node -- the xml node (xml.dom.minidom.Node) """ term1_attrib = node.getAttribute('left-field') term1_value = node.getAttribute('left-value') op = node.nodeName.lower() term2_attrib = node.getAttribute('right-field') term2_value = node.getAttribute('right-value') if op not in _op_map: _exc('Invalid operator') if term1_attrib != '' and term1_value != '': _exc('Both, left-field and left-value attributes found') elif term1_attrib == '' and term1_value == '': _exc('left-field or left-value attribute required') elif term1_value != '': left = term1_value else: left = operators.Attrib(term1_attrib) if term2_attrib != '' and term2_value != '': _exc('Both, right-field and right-value attributes found') elif term2_attrib == '' and term2_value == '': _exc('right-field or right-value attribute required') elif term2_value != '': right = term2_value else: right = operators.Attrib(term2_attrib) return _op_map[op](left, right)
[ "Reads", "the", "logical", "tag", "from", "the", "given", "node", "returns", "a", "Condition", "object", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L100-L129
[ "def", "deserialize_logical", "(", "self", ",", "node", ")", ":", "term1_attrib", "=", "node", ".", "getAttribute", "(", "'left-field'", ")", "term1_value", "=", "node", ".", "getAttribute", "(", "'left-value'", ")", "op", "=", "node", ".", "nodeName", ".", "lower", "(", ")", "term2_attrib", "=", "node", ".", "getAttribute", "(", "'right-field'", ")", "term2_value", "=", "node", ".", "getAttribute", "(", "'right-value'", ")", "if", "op", "not", "in", "_op_map", ":", "_exc", "(", "'Invalid operator'", ")", "if", "term1_attrib", "!=", "''", "and", "term1_value", "!=", "''", ":", "_exc", "(", "'Both, left-field and left-value attributes found'", ")", "elif", "term1_attrib", "==", "''", "and", "term1_value", "==", "''", ":", "_exc", "(", "'left-field or left-value attribute required'", ")", "elif", "term1_value", "!=", "''", ":", "left", "=", "term1_value", "else", ":", "left", "=", "operators", ".", "Attrib", "(", "term1_attrib", ")", "if", "term2_attrib", "!=", "''", "and", "term2_value", "!=", "''", ":", "_exc", "(", "'Both, right-field and right-value attributes found'", ")", "elif", "term2_attrib", "==", "''", "and", "term2_value", "==", "''", ":", "_exc", "(", "'right-field or right-value attribute required'", ")", "elif", "term2_value", "!=", "''", ":", "right", "=", "term2_value", "else", ":", "right", "=", "operators", ".", "Attrib", "(", "term2_attrib", ")", "return", "_op_map", "[", "op", "]", "(", "left", ",", "right", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_condition
Reads the conditional statement from the given node. workflow -- the workflow with which the concurrence is associated start_node -- the xml structure (xml.dom.minidom.Node)
SpiffWorkflow/serializer/prettyxml.py
def deserialize_condition(self, workflow, start_node): """ Reads the conditional statement from the given node. workflow -- the workflow with which the concurrence is associated start_node -- the xml structure (xml.dom.minidom.Node) """ # Collect all information. condition = None spec_name = None for node in start_node.childNodes: if node.nodeType != minidom.Node.ELEMENT_NODE: continue if node.nodeName.lower() == 'successor': if spec_name is not None: _exc('Duplicate task name %s' % spec_name) if node.firstChild is None: _exc('Successor tag without a task name') spec_name = node.firstChild.nodeValue elif node.nodeName.lower() in _op_map: if condition is not None: _exc('Multiple conditions are not yet supported') condition = self.deserialize_logical(node) else: _exc('Unknown node: %s' % node.nodeName) if condition is None: _exc('Missing condition in conditional statement') if spec_name is None: _exc('A %s has no task specified' % start_node.nodeName) return condition, spec_name
def deserialize_condition(self, workflow, start_node): """ Reads the conditional statement from the given node. workflow -- the workflow with which the concurrence is associated start_node -- the xml structure (xml.dom.minidom.Node) """ # Collect all information. condition = None spec_name = None for node in start_node.childNodes: if node.nodeType != minidom.Node.ELEMENT_NODE: continue if node.nodeName.lower() == 'successor': if spec_name is not None: _exc('Duplicate task name %s' % spec_name) if node.firstChild is None: _exc('Successor tag without a task name') spec_name = node.firstChild.nodeValue elif node.nodeName.lower() in _op_map: if condition is not None: _exc('Multiple conditions are not yet supported') condition = self.deserialize_logical(node) else: _exc('Unknown node: %s' % node.nodeName) if condition is None: _exc('Missing condition in conditional statement') if spec_name is None: _exc('A %s has no task specified' % start_node.nodeName) return condition, spec_name
[ "Reads", "the", "conditional", "statement", "from", "the", "given", "node", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L131-L161
[ "def", "deserialize_condition", "(", "self", ",", "workflow", ",", "start_node", ")", ":", "# Collect all information.", "condition", "=", "None", "spec_name", "=", "None", "for", "node", "in", "start_node", ".", "childNodes", ":", "if", "node", ".", "nodeType", "!=", "minidom", ".", "Node", ".", "ELEMENT_NODE", ":", "continue", "if", "node", ".", "nodeName", ".", "lower", "(", ")", "==", "'successor'", ":", "if", "spec_name", "is", "not", "None", ":", "_exc", "(", "'Duplicate task name %s'", "%", "spec_name", ")", "if", "node", ".", "firstChild", "is", "None", ":", "_exc", "(", "'Successor tag without a task name'", ")", "spec_name", "=", "node", ".", "firstChild", ".", "nodeValue", "elif", "node", ".", "nodeName", ".", "lower", "(", ")", "in", "_op_map", ":", "if", "condition", "is", "not", "None", ":", "_exc", "(", "'Multiple conditions are not yet supported'", ")", "condition", "=", "self", ".", "deserialize_logical", "(", "node", ")", "else", ":", "_exc", "(", "'Unknown node: %s'", "%", "node", ".", "nodeName", ")", "if", "condition", "is", "None", ":", "_exc", "(", "'Missing condition in conditional statement'", ")", "if", "spec_name", "is", "None", ":", "_exc", "(", "'A %s has no task specified'", "%", "start_node", ".", "nodeName", ")", "return", "condition", ",", "spec_name" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_task_spec
Reads the task from the given node and returns a tuple (start, end) that contains the stream of objects that model the behavior. workflow -- the workflow with which the task is associated start_node -- the xml structure (xml.dom.minidom.Node)
SpiffWorkflow/serializer/prettyxml.py
def deserialize_task_spec(self, workflow, start_node, read_specs): """ Reads the task from the given node and returns a tuple (start, end) that contains the stream of objects that model the behavior. workflow -- the workflow with which the task is associated start_node -- the xml structure (xml.dom.minidom.Node) """ # Extract attributes from the node. nodetype = start_node.nodeName.lower() name = start_node.getAttribute('name').lower() context = start_node.getAttribute('context').lower() mutex = start_node.getAttribute('mutex').lower() cancel = start_node.getAttribute('cancel').lower() success = start_node.getAttribute('success').lower() times = start_node.getAttribute('times').lower() times_field = start_node.getAttribute('times-field').lower() threshold = start_node.getAttribute('threshold').lower() threshold_field = start_node.getAttribute('threshold-field').lower() file = start_node.getAttribute('file').lower() file_field = start_node.getAttribute('file-field').lower() kwargs = {'lock': [], 'data': {}, 'defines': {}, 'pre_assign': [], 'post_assign': []} if nodetype not in _spec_map: _exc('Invalid task type "%s"' % nodetype) if nodetype == 'start-task': name = 'start' if name == '': _exc('Invalid task name "%s"' % name) if name in read_specs: _exc('Duplicate task name "%s"' % name) if cancel != '' and cancel != '0': kwargs['cancel'] = True if success != '' and success != '0': kwargs['success'] = True if times != '': kwargs['times'] = int(times) if times_field != '': kwargs['times'] = operators.Attrib(times_field) if threshold != '': kwargs['threshold'] = int(threshold) if threshold_field != '': kwargs['threshold'] = operators.Attrib(threshold_field) if file != '': kwargs['file'] = file if file_field != '': kwargs['file'] = operators.Attrib(file_field) if nodetype == 'choose': kwargs['choice'] = [] if nodetype == 'trigger': context = [context] if mutex != '': context = mutex # Walk through the children of the node. successors = [] for node in start_node.childNodes: if node.nodeType != minidom.Node.ELEMENT_NODE: continue if node.nodeName == 'description': kwargs['description'] = node.firstChild.nodeValue elif node.nodeName == 'successor' \ or node.nodeName == 'default-successor': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) successors.append((None, node.firstChild.nodeValue)) elif node.nodeName == 'conditional-successor': successors.append(self.deserialize_condition(workflow, node)) elif node.nodeName == 'define': key, value = self.deserialize_data(workflow, node) kwargs['defines'][key] = value # "property" tag exists for backward compatibility. elif node.nodeName == 'data' or node.nodeName == 'property': key, value = self.deserialize_data(workflow, node) kwargs['data'][key] = value elif node.nodeName == 'pre-assign': kwargs['pre_assign'].append( self.deserialize_assign(workflow, node)) elif node.nodeName == 'post-assign': kwargs['post_assign'].append( self.deserialize_assign(workflow, node)) elif node.nodeName == 'in': kwargs['in_assign'] = self.deserialize_assign_list( workflow, node) elif node.nodeName == 'out': kwargs['out_assign'] = self.deserialize_assign_list( workflow, node) elif node.nodeName == 'cancel': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) if context == '': context = [] elif not isinstance(context, list): context = [context] context.append(node.firstChild.nodeValue) elif node.nodeName == 'lock': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) kwargs['lock'].append(node.firstChild.nodeValue) elif node.nodeName == 'pick': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) kwargs['choice'].append(node.firstChild.nodeValue) else: _exc('Unknown node: %s' % node.nodeName) # Create a new instance of the task spec. module = _spec_map[nodetype] if nodetype == 'start-task': spec = module(workflow, **kwargs) elif nodetype == 'multi-instance' or nodetype == 'thread-split': if times == '' and times_field == '': _exc('Missing "times" or "times-field" in "%s"' % name) elif times != '' and times_field != '': _exc('Both, "times" and "times-field" in "%s"' % name) spec = module(workflow, name, **kwargs) elif context == '': spec = module(workflow, name, **kwargs) else: spec = module(workflow, name, context, **kwargs) read_specs[name] = spec, successors
def deserialize_task_spec(self, workflow, start_node, read_specs): """ Reads the task from the given node and returns a tuple (start, end) that contains the stream of objects that model the behavior. workflow -- the workflow with which the task is associated start_node -- the xml structure (xml.dom.minidom.Node) """ # Extract attributes from the node. nodetype = start_node.nodeName.lower() name = start_node.getAttribute('name').lower() context = start_node.getAttribute('context').lower() mutex = start_node.getAttribute('mutex').lower() cancel = start_node.getAttribute('cancel').lower() success = start_node.getAttribute('success').lower() times = start_node.getAttribute('times').lower() times_field = start_node.getAttribute('times-field').lower() threshold = start_node.getAttribute('threshold').lower() threshold_field = start_node.getAttribute('threshold-field').lower() file = start_node.getAttribute('file').lower() file_field = start_node.getAttribute('file-field').lower() kwargs = {'lock': [], 'data': {}, 'defines': {}, 'pre_assign': [], 'post_assign': []} if nodetype not in _spec_map: _exc('Invalid task type "%s"' % nodetype) if nodetype == 'start-task': name = 'start' if name == '': _exc('Invalid task name "%s"' % name) if name in read_specs: _exc('Duplicate task name "%s"' % name) if cancel != '' and cancel != '0': kwargs['cancel'] = True if success != '' and success != '0': kwargs['success'] = True if times != '': kwargs['times'] = int(times) if times_field != '': kwargs['times'] = operators.Attrib(times_field) if threshold != '': kwargs['threshold'] = int(threshold) if threshold_field != '': kwargs['threshold'] = operators.Attrib(threshold_field) if file != '': kwargs['file'] = file if file_field != '': kwargs['file'] = operators.Attrib(file_field) if nodetype == 'choose': kwargs['choice'] = [] if nodetype == 'trigger': context = [context] if mutex != '': context = mutex # Walk through the children of the node. successors = [] for node in start_node.childNodes: if node.nodeType != minidom.Node.ELEMENT_NODE: continue if node.nodeName == 'description': kwargs['description'] = node.firstChild.nodeValue elif node.nodeName == 'successor' \ or node.nodeName == 'default-successor': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) successors.append((None, node.firstChild.nodeValue)) elif node.nodeName == 'conditional-successor': successors.append(self.deserialize_condition(workflow, node)) elif node.nodeName == 'define': key, value = self.deserialize_data(workflow, node) kwargs['defines'][key] = value # "property" tag exists for backward compatibility. elif node.nodeName == 'data' or node.nodeName == 'property': key, value = self.deserialize_data(workflow, node) kwargs['data'][key] = value elif node.nodeName == 'pre-assign': kwargs['pre_assign'].append( self.deserialize_assign(workflow, node)) elif node.nodeName == 'post-assign': kwargs['post_assign'].append( self.deserialize_assign(workflow, node)) elif node.nodeName == 'in': kwargs['in_assign'] = self.deserialize_assign_list( workflow, node) elif node.nodeName == 'out': kwargs['out_assign'] = self.deserialize_assign_list( workflow, node) elif node.nodeName == 'cancel': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) if context == '': context = [] elif not isinstance(context, list): context = [context] context.append(node.firstChild.nodeValue) elif node.nodeName == 'lock': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) kwargs['lock'].append(node.firstChild.nodeValue) elif node.nodeName == 'pick': if node.firstChild is None: _exc('Empty %s tag' % node.nodeName) kwargs['choice'].append(node.firstChild.nodeValue) else: _exc('Unknown node: %s' % node.nodeName) # Create a new instance of the task spec. module = _spec_map[nodetype] if nodetype == 'start-task': spec = module(workflow, **kwargs) elif nodetype == 'multi-instance' or nodetype == 'thread-split': if times == '' and times_field == '': _exc('Missing "times" or "times-field" in "%s"' % name) elif times != '' and times_field != '': _exc('Both, "times" and "times-field" in "%s"' % name) spec = module(workflow, name, **kwargs) elif context == '': spec = module(workflow, name, **kwargs) else: spec = module(workflow, name, context, **kwargs) read_specs[name] = spec, successors
[ "Reads", "the", "task", "from", "the", "given", "node", "and", "returns", "a", "tuple", "(", "start", "end", ")", "that", "contains", "the", "stream", "of", "objects", "that", "model", "the", "behavior", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L163-L288
[ "def", "deserialize_task_spec", "(", "self", ",", "workflow", ",", "start_node", ",", "read_specs", ")", ":", "# Extract attributes from the node.", "nodetype", "=", "start_node", ".", "nodeName", ".", "lower", "(", ")", "name", "=", "start_node", ".", "getAttribute", "(", "'name'", ")", ".", "lower", "(", ")", "context", "=", "start_node", ".", "getAttribute", "(", "'context'", ")", ".", "lower", "(", ")", "mutex", "=", "start_node", ".", "getAttribute", "(", "'mutex'", ")", ".", "lower", "(", ")", "cancel", "=", "start_node", ".", "getAttribute", "(", "'cancel'", ")", ".", "lower", "(", ")", "success", "=", "start_node", ".", "getAttribute", "(", "'success'", ")", ".", "lower", "(", ")", "times", "=", "start_node", ".", "getAttribute", "(", "'times'", ")", ".", "lower", "(", ")", "times_field", "=", "start_node", ".", "getAttribute", "(", "'times-field'", ")", ".", "lower", "(", ")", "threshold", "=", "start_node", ".", "getAttribute", "(", "'threshold'", ")", ".", "lower", "(", ")", "threshold_field", "=", "start_node", ".", "getAttribute", "(", "'threshold-field'", ")", ".", "lower", "(", ")", "file", "=", "start_node", ".", "getAttribute", "(", "'file'", ")", ".", "lower", "(", ")", "file_field", "=", "start_node", ".", "getAttribute", "(", "'file-field'", ")", ".", "lower", "(", ")", "kwargs", "=", "{", "'lock'", ":", "[", "]", ",", "'data'", ":", "{", "}", ",", "'defines'", ":", "{", "}", ",", "'pre_assign'", ":", "[", "]", ",", "'post_assign'", ":", "[", "]", "}", "if", "nodetype", "not", "in", "_spec_map", ":", "_exc", "(", "'Invalid task type \"%s\"'", "%", "nodetype", ")", "if", "nodetype", "==", "'start-task'", ":", "name", "=", "'start'", "if", "name", "==", "''", ":", "_exc", "(", "'Invalid task name \"%s\"'", "%", "name", ")", "if", "name", "in", "read_specs", ":", "_exc", "(", "'Duplicate task name \"%s\"'", "%", "name", ")", "if", "cancel", "!=", "''", "and", "cancel", "!=", "'0'", ":", "kwargs", "[", "'cancel'", "]", "=", "True", "if", "success", "!=", "''", "and", "success", "!=", "'0'", ":", "kwargs", "[", "'success'", "]", "=", "True", "if", "times", "!=", "''", ":", "kwargs", "[", "'times'", "]", "=", "int", "(", "times", ")", "if", "times_field", "!=", "''", ":", "kwargs", "[", "'times'", "]", "=", "operators", ".", "Attrib", "(", "times_field", ")", "if", "threshold", "!=", "''", ":", "kwargs", "[", "'threshold'", "]", "=", "int", "(", "threshold", ")", "if", "threshold_field", "!=", "''", ":", "kwargs", "[", "'threshold'", "]", "=", "operators", ".", "Attrib", "(", "threshold_field", ")", "if", "file", "!=", "''", ":", "kwargs", "[", "'file'", "]", "=", "file", "if", "file_field", "!=", "''", ":", "kwargs", "[", "'file'", "]", "=", "operators", ".", "Attrib", "(", "file_field", ")", "if", "nodetype", "==", "'choose'", ":", "kwargs", "[", "'choice'", "]", "=", "[", "]", "if", "nodetype", "==", "'trigger'", ":", "context", "=", "[", "context", "]", "if", "mutex", "!=", "''", ":", "context", "=", "mutex", "# Walk through the children of the node.", "successors", "=", "[", "]", "for", "node", "in", "start_node", ".", "childNodes", ":", "if", "node", ".", "nodeType", "!=", "minidom", ".", "Node", ".", "ELEMENT_NODE", ":", "continue", "if", "node", ".", "nodeName", "==", "'description'", ":", "kwargs", "[", "'description'", "]", "=", "node", ".", "firstChild", ".", "nodeValue", "elif", "node", ".", "nodeName", "==", "'successor'", "or", "node", ".", "nodeName", "==", "'default-successor'", ":", "if", "node", ".", "firstChild", "is", "None", ":", "_exc", "(", "'Empty %s tag'", "%", "node", ".", "nodeName", ")", "successors", ".", "append", "(", "(", "None", ",", "node", ".", "firstChild", ".", "nodeValue", ")", ")", "elif", "node", ".", "nodeName", "==", "'conditional-successor'", ":", "successors", ".", "append", "(", "self", ".", "deserialize_condition", "(", "workflow", ",", "node", ")", ")", "elif", "node", ".", "nodeName", "==", "'define'", ":", "key", ",", "value", "=", "self", ".", "deserialize_data", "(", "workflow", ",", "node", ")", "kwargs", "[", "'defines'", "]", "[", "key", "]", "=", "value", "# \"property\" tag exists for backward compatibility.", "elif", "node", ".", "nodeName", "==", "'data'", "or", "node", ".", "nodeName", "==", "'property'", ":", "key", ",", "value", "=", "self", ".", "deserialize_data", "(", "workflow", ",", "node", ")", "kwargs", "[", "'data'", "]", "[", "key", "]", "=", "value", "elif", "node", ".", "nodeName", "==", "'pre-assign'", ":", "kwargs", "[", "'pre_assign'", "]", ".", "append", "(", "self", ".", "deserialize_assign", "(", "workflow", ",", "node", ")", ")", "elif", "node", ".", "nodeName", "==", "'post-assign'", ":", "kwargs", "[", "'post_assign'", "]", ".", "append", "(", "self", ".", "deserialize_assign", "(", "workflow", ",", "node", ")", ")", "elif", "node", ".", "nodeName", "==", "'in'", ":", "kwargs", "[", "'in_assign'", "]", "=", "self", ".", "deserialize_assign_list", "(", "workflow", ",", "node", ")", "elif", "node", ".", "nodeName", "==", "'out'", ":", "kwargs", "[", "'out_assign'", "]", "=", "self", ".", "deserialize_assign_list", "(", "workflow", ",", "node", ")", "elif", "node", ".", "nodeName", "==", "'cancel'", ":", "if", "node", ".", "firstChild", "is", "None", ":", "_exc", "(", "'Empty %s tag'", "%", "node", ".", "nodeName", ")", "if", "context", "==", "''", ":", "context", "=", "[", "]", "elif", "not", "isinstance", "(", "context", ",", "list", ")", ":", "context", "=", "[", "context", "]", "context", ".", "append", "(", "node", ".", "firstChild", ".", "nodeValue", ")", "elif", "node", ".", "nodeName", "==", "'lock'", ":", "if", "node", ".", "firstChild", "is", "None", ":", "_exc", "(", "'Empty %s tag'", "%", "node", ".", "nodeName", ")", "kwargs", "[", "'lock'", "]", ".", "append", "(", "node", ".", "firstChild", ".", "nodeValue", ")", "elif", "node", ".", "nodeName", "==", "'pick'", ":", "if", "node", ".", "firstChild", "is", "None", ":", "_exc", "(", "'Empty %s tag'", "%", "node", ".", "nodeName", ")", "kwargs", "[", "'choice'", "]", ".", "append", "(", "node", ".", "firstChild", ".", "nodeValue", ")", "else", ":", "_exc", "(", "'Unknown node: %s'", "%", "node", ".", "nodeName", ")", "# Create a new instance of the task spec.", "module", "=", "_spec_map", "[", "nodetype", "]", "if", "nodetype", "==", "'start-task'", ":", "spec", "=", "module", "(", "workflow", ",", "*", "*", "kwargs", ")", "elif", "nodetype", "==", "'multi-instance'", "or", "nodetype", "==", "'thread-split'", ":", "if", "times", "==", "''", "and", "times_field", "==", "''", ":", "_exc", "(", "'Missing \"times\" or \"times-field\" in \"%s\"'", "%", "name", ")", "elif", "times", "!=", "''", "and", "times_field", "!=", "''", ":", "_exc", "(", "'Both, \"times\" and \"times-field\" in \"%s\"'", "%", "name", ")", "spec", "=", "module", "(", "workflow", ",", "name", ",", "*", "*", "kwargs", ")", "elif", "context", "==", "''", ":", "spec", "=", "module", "(", "workflow", ",", "name", ",", "*", "*", "kwargs", ")", "else", ":", "spec", "=", "module", "(", "workflow", ",", "name", ",", "context", ",", "*", "*", "kwargs", ")", "read_specs", "[", "name", "]", "=", "spec", ",", "successors" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
XmlSerializer.deserialize_workflow_spec
Reads the workflow from the given XML structure and returns a WorkflowSpec instance.
SpiffWorkflow/serializer/prettyxml.py
def deserialize_workflow_spec(self, s_state, filename=None): """ Reads the workflow from the given XML structure and returns a WorkflowSpec instance. """ dom = minidom.parseString(s_state) node = dom.getElementsByTagName('process-definition')[0] name = node.getAttribute('name') if name == '': _exc('%s without a name attribute' % node.nodeName) # Read all task specs and create a list of successors. workflow_spec = specs.WorkflowSpec(name, filename) del workflow_spec.task_specs['Start'] end = specs.Simple(workflow_spec, 'End'), [] read_specs = dict(end=end) for child_node in node.childNodes: if child_node.nodeType != minidom.Node.ELEMENT_NODE: continue if child_node.nodeName == 'name': workflow_spec.name = child_node.firstChild.nodeValue elif child_node.nodeName == 'description': workflow_spec.description = child_node.firstChild.nodeValue elif child_node.nodeName.lower() in _spec_map: self.deserialize_task_spec( workflow_spec, child_node, read_specs) else: _exc('Unknown node: %s' % child_node.nodeName) # Remove the default start-task from the workflow. workflow_spec.start = read_specs['start'][0] # Connect all task specs. for name in read_specs: spec, successors = read_specs[name] for condition, successor_name in successors: if successor_name not in read_specs: _exc('Unknown successor: "%s"' % successor_name) successor, foo = read_specs[successor_name] if condition is None: spec.connect(successor) else: spec.connect_if(condition, successor) return workflow_spec
def deserialize_workflow_spec(self, s_state, filename=None): """ Reads the workflow from the given XML structure and returns a WorkflowSpec instance. """ dom = minidom.parseString(s_state) node = dom.getElementsByTagName('process-definition')[0] name = node.getAttribute('name') if name == '': _exc('%s without a name attribute' % node.nodeName) # Read all task specs and create a list of successors. workflow_spec = specs.WorkflowSpec(name, filename) del workflow_spec.task_specs['Start'] end = specs.Simple(workflow_spec, 'End'), [] read_specs = dict(end=end) for child_node in node.childNodes: if child_node.nodeType != minidom.Node.ELEMENT_NODE: continue if child_node.nodeName == 'name': workflow_spec.name = child_node.firstChild.nodeValue elif child_node.nodeName == 'description': workflow_spec.description = child_node.firstChild.nodeValue elif child_node.nodeName.lower() in _spec_map: self.deserialize_task_spec( workflow_spec, child_node, read_specs) else: _exc('Unknown node: %s' % child_node.nodeName) # Remove the default start-task from the workflow. workflow_spec.start = read_specs['start'][0] # Connect all task specs. for name in read_specs: spec, successors = read_specs[name] for condition, successor_name in successors: if successor_name not in read_specs: _exc('Unknown successor: "%s"' % successor_name) successor, foo = read_specs[successor_name] if condition is None: spec.connect(successor) else: spec.connect_if(condition, successor) return workflow_spec
[ "Reads", "the", "workflow", "from", "the", "given", "XML", "structure", "and", "returns", "a", "WorkflowSpec", "instance", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/serializer/prettyxml.py#L290-L333
[ "def", "deserialize_workflow_spec", "(", "self", ",", "s_state", ",", "filename", "=", "None", ")", ":", "dom", "=", "minidom", ".", "parseString", "(", "s_state", ")", "node", "=", "dom", ".", "getElementsByTagName", "(", "'process-definition'", ")", "[", "0", "]", "name", "=", "node", ".", "getAttribute", "(", "'name'", ")", "if", "name", "==", "''", ":", "_exc", "(", "'%s without a name attribute'", "%", "node", ".", "nodeName", ")", "# Read all task specs and create a list of successors.", "workflow_spec", "=", "specs", ".", "WorkflowSpec", "(", "name", ",", "filename", ")", "del", "workflow_spec", ".", "task_specs", "[", "'Start'", "]", "end", "=", "specs", ".", "Simple", "(", "workflow_spec", ",", "'End'", ")", ",", "[", "]", "read_specs", "=", "dict", "(", "end", "=", "end", ")", "for", "child_node", "in", "node", ".", "childNodes", ":", "if", "child_node", ".", "nodeType", "!=", "minidom", ".", "Node", ".", "ELEMENT_NODE", ":", "continue", "if", "child_node", ".", "nodeName", "==", "'name'", ":", "workflow_spec", ".", "name", "=", "child_node", ".", "firstChild", ".", "nodeValue", "elif", "child_node", ".", "nodeName", "==", "'description'", ":", "workflow_spec", ".", "description", "=", "child_node", ".", "firstChild", ".", "nodeValue", "elif", "child_node", ".", "nodeName", ".", "lower", "(", ")", "in", "_spec_map", ":", "self", ".", "deserialize_task_spec", "(", "workflow_spec", ",", "child_node", ",", "read_specs", ")", "else", ":", "_exc", "(", "'Unknown node: %s'", "%", "child_node", ".", "nodeName", ")", "# Remove the default start-task from the workflow.", "workflow_spec", ".", "start", "=", "read_specs", "[", "'start'", "]", "[", "0", "]", "# Connect all task specs.", "for", "name", "in", "read_specs", ":", "spec", ",", "successors", "=", "read_specs", "[", "name", "]", "for", "condition", ",", "successor_name", "in", "successors", ":", "if", "successor_name", "not", "in", "read_specs", ":", "_exc", "(", "'Unknown successor: \"%s\"'", "%", "successor_name", ")", "successor", ",", "foo", "=", "read_specs", "[", "successor_name", "]", "if", "condition", "is", "None", ":", "spec", ".", "connect", "(", "successor", ")", "else", ":", "spec", ".", "connect_if", "(", "condition", ",", "successor", ")", "return", "workflow_spec" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
TimerEventDefinition.has_fired
The Timer is considered to have fired if the evaluated dateTime expression is before datetime.datetime.now()
SpiffWorkflow/bpmn/specs/event_definitions.py
def has_fired(self, my_task): """ The Timer is considered to have fired if the evaluated dateTime expression is before datetime.datetime.now() """ dt = my_task.workflow.script_engine.evaluate(my_task, self.dateTime) if dt is None: return False if dt.tzinfo: tz = dt.tzinfo now = tz.fromutc(datetime.datetime.utcnow().replace(tzinfo=tz)) else: now = datetime.datetime.now() return now > dt
def has_fired(self, my_task): """ The Timer is considered to have fired if the evaluated dateTime expression is before datetime.datetime.now() """ dt = my_task.workflow.script_engine.evaluate(my_task, self.dateTime) if dt is None: return False if dt.tzinfo: tz = dt.tzinfo now = tz.fromutc(datetime.datetime.utcnow().replace(tzinfo=tz)) else: now = datetime.datetime.now() return now > dt
[ "The", "Timer", "is", "considered", "to", "have", "fired", "if", "the", "evaluated", "dateTime", "expression", "is", "before", "datetime", ".", "datetime", ".", "now", "()" ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/specs/event_definitions.py#L98-L111
[ "def", "has_fired", "(", "self", ",", "my_task", ")", ":", "dt", "=", "my_task", ".", "workflow", ".", "script_engine", ".", "evaluate", "(", "my_task", ",", "self", ".", "dateTime", ")", "if", "dt", "is", "None", ":", "return", "False", "if", "dt", ".", "tzinfo", ":", "tz", "=", "dt", ".", "tzinfo", "now", "=", "tz", ".", "fromutc", "(", "datetime", ".", "datetime", ".", "utcnow", "(", ")", ".", "replace", "(", "tzinfo", "=", "tz", ")", ")", "else", ":", "now", "=", "datetime", ".", "datetime", ".", "now", "(", ")", "return", "now", ">", "dt" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
WorkflowSpec._add_notify
Called by a task spec when it was added into the workflow.
SpiffWorkflow/specs/WorkflowSpec.py
def _add_notify(self, task_spec): """ Called by a task spec when it was added into the workflow. """ if task_spec.name in self.task_specs: raise KeyError('Duplicate task spec name: ' + task_spec.name) self.task_specs[task_spec.name] = task_spec task_spec.id = len(self.task_specs)
def _add_notify(self, task_spec): """ Called by a task spec when it was added into the workflow. """ if task_spec.name in self.task_specs: raise KeyError('Duplicate task spec name: ' + task_spec.name) self.task_specs[task_spec.name] = task_spec task_spec.id = len(self.task_specs)
[ "Called", "by", "a", "task", "spec", "when", "it", "was", "added", "into", "the", "workflow", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/WorkflowSpec.py#L47-L54
[ "def", "_add_notify", "(", "self", ",", "task_spec", ")", ":", "if", "task_spec", ".", "name", "in", "self", ".", "task_specs", ":", "raise", "KeyError", "(", "'Duplicate task spec name: '", "+", "task_spec", ".", "name", ")", "self", ".", "task_specs", "[", "task_spec", ".", "name", "]", "=", "task_spec", "task_spec", ".", "id", "=", "len", "(", "self", ".", "task_specs", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
WorkflowSpec.validate
Checks integrity of workflow and reports any problems with it. Detects: - loops (tasks that wait on each other in a loop) :returns: empty list if valid, a list of errors if not
SpiffWorkflow/specs/WorkflowSpec.py
def validate(self): """Checks integrity of workflow and reports any problems with it. Detects: - loops (tasks that wait on each other in a loop) :returns: empty list if valid, a list of errors if not """ results = [] from ..specs import Join def recursive_find_loop(task, history): current = history[:] current.append(task) if isinstance(task, Join): if task in history: msg = "Found loop with '%s': %s then '%s' again" % ( task.name, '->'.join([p.name for p in history]), task.name) raise Exception(msg) for predecessor in task.inputs: recursive_find_loop(predecessor, current) for parent in task.inputs: recursive_find_loop(parent, current) for task_id, task in list(self.task_specs.items()): # Check for cyclic waits try: recursive_find_loop(task, []) except Exception as exc: results.append(exc.__str__()) # Check for disconnected tasks if not task.inputs and task.name not in ['Start', 'Root']: if task.outputs: results.append("Task '%s' is disconnected (no inputs)" % task.name) else: LOG.debug("Task '%s' is not being used" % task.name) return results
def validate(self): """Checks integrity of workflow and reports any problems with it. Detects: - loops (tasks that wait on each other in a loop) :returns: empty list if valid, a list of errors if not """ results = [] from ..specs import Join def recursive_find_loop(task, history): current = history[:] current.append(task) if isinstance(task, Join): if task in history: msg = "Found loop with '%s': %s then '%s' again" % ( task.name, '->'.join([p.name for p in history]), task.name) raise Exception(msg) for predecessor in task.inputs: recursive_find_loop(predecessor, current) for parent in task.inputs: recursive_find_loop(parent, current) for task_id, task in list(self.task_specs.items()): # Check for cyclic waits try: recursive_find_loop(task, []) except Exception as exc: results.append(exc.__str__()) # Check for disconnected tasks if not task.inputs and task.name not in ['Start', 'Root']: if task.outputs: results.append("Task '%s' is disconnected (no inputs)" % task.name) else: LOG.debug("Task '%s' is not being used" % task.name) return results
[ "Checks", "integrity", "of", "workflow", "and", "reports", "any", "problems", "with", "it", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/WorkflowSpec.py#L67-L107
[ "def", "validate", "(", "self", ")", ":", "results", "=", "[", "]", "from", ".", ".", "specs", "import", "Join", "def", "recursive_find_loop", "(", "task", ",", "history", ")", ":", "current", "=", "history", "[", ":", "]", "current", ".", "append", "(", "task", ")", "if", "isinstance", "(", "task", ",", "Join", ")", ":", "if", "task", "in", "history", ":", "msg", "=", "\"Found loop with '%s': %s then '%s' again\"", "%", "(", "task", ".", "name", ",", "'->'", ".", "join", "(", "[", "p", ".", "name", "for", "p", "in", "history", "]", ")", ",", "task", ".", "name", ")", "raise", "Exception", "(", "msg", ")", "for", "predecessor", "in", "task", ".", "inputs", ":", "recursive_find_loop", "(", "predecessor", ",", "current", ")", "for", "parent", "in", "task", ".", "inputs", ":", "recursive_find_loop", "(", "parent", ",", "current", ")", "for", "task_id", ",", "task", "in", "list", "(", "self", ".", "task_specs", ".", "items", "(", ")", ")", ":", "# Check for cyclic waits", "try", ":", "recursive_find_loop", "(", "task", ",", "[", "]", ")", "except", "Exception", "as", "exc", ":", "results", ".", "append", "(", "exc", ".", "__str__", "(", ")", ")", "# Check for disconnected tasks", "if", "not", "task", ".", "inputs", "and", "task", ".", "name", "not", "in", "[", "'Start'", ",", "'Root'", "]", ":", "if", "task", ".", "outputs", ":", "results", ".", "append", "(", "\"Task '%s' is disconnected (no inputs)\"", "%", "task", ".", "name", ")", "else", ":", "LOG", ".", "debug", "(", "\"Task '%s' is not being used\"", "%", "task", ".", "name", ")", "return", "results" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BpmnWorkflow.accept_message
Indicate to the workflow that a message has been received. The message will be processed by any waiting Intermediate or Boundary Message Events, that are waiting for the message.
SpiffWorkflow/bpmn/workflow.py
def accept_message(self, message): """ Indicate to the workflow that a message has been received. The message will be processed by any waiting Intermediate or Boundary Message Events, that are waiting for the message. """ assert not self.read_only self.refresh_waiting_tasks() self.do_engine_steps() for my_task in Task.Iterator(self.task_tree, Task.WAITING): my_task.task_spec.accept_message(my_task, message)
def accept_message(self, message): """ Indicate to the workflow that a message has been received. The message will be processed by any waiting Intermediate or Boundary Message Events, that are waiting for the message. """ assert not self.read_only self.refresh_waiting_tasks() self.do_engine_steps() for my_task in Task.Iterator(self.task_tree, Task.WAITING): my_task.task_spec.accept_message(my_task, message)
[ "Indicate", "to", "the", "workflow", "that", "a", "message", "has", "been", "received", ".", "The", "message", "will", "be", "processed", "by", "any", "waiting", "Intermediate", "or", "Boundary", "Message", "Events", "that", "are", "waiting", "for", "the", "message", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/workflow.py#L51-L61
[ "def", "accept_message", "(", "self", ",", "message", ")", ":", "assert", "not", "self", ".", "read_only", "self", ".", "refresh_waiting_tasks", "(", ")", "self", ".", "do_engine_steps", "(", ")", "for", "my_task", "in", "Task", ".", "Iterator", "(", "self", ".", "task_tree", ",", "Task", ".", "WAITING", ")", ":", "my_task", ".", "task_spec", ".", "accept_message", "(", "my_task", ",", "message", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BpmnWorkflow.do_engine_steps
Execute any READY tasks that are engine specific (for example, gateways or script tasks). This is done in a loop, so it will keep completing those tasks until there are only READY User tasks, or WAITING tasks left.
SpiffWorkflow/bpmn/workflow.py
def do_engine_steps(self): """ Execute any READY tasks that are engine specific (for example, gateways or script tasks). This is done in a loop, so it will keep completing those tasks until there are only READY User tasks, or WAITING tasks left. """ assert not self.read_only engine_steps = list( [t for t in self.get_tasks(Task.READY) if self._is_engine_task(t.task_spec)]) while engine_steps: for task in engine_steps: task.complete() engine_steps = list( [t for t in self.get_tasks(Task.READY) if self._is_engine_task(t.task_spec)])
def do_engine_steps(self): """ Execute any READY tasks that are engine specific (for example, gateways or script tasks). This is done in a loop, so it will keep completing those tasks until there are only READY User tasks, or WAITING tasks left. """ assert not self.read_only engine_steps = list( [t for t in self.get_tasks(Task.READY) if self._is_engine_task(t.task_spec)]) while engine_steps: for task in engine_steps: task.complete() engine_steps = list( [t for t in self.get_tasks(Task.READY) if self._is_engine_task(t.task_spec)])
[ "Execute", "any", "READY", "tasks", "that", "are", "engine", "specific", "(", "for", "example", "gateways", "or", "script", "tasks", ")", ".", "This", "is", "done", "in", "a", "loop", "so", "it", "will", "keep", "completing", "those", "tasks", "until", "there", "are", "only", "READY", "User", "tasks", "or", "WAITING", "tasks", "left", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/workflow.py#L63-L79
[ "def", "do_engine_steps", "(", "self", ")", ":", "assert", "not", "self", ".", "read_only", "engine_steps", "=", "list", "(", "[", "t", "for", "t", "in", "self", ".", "get_tasks", "(", "Task", ".", "READY", ")", "if", "self", ".", "_is_engine_task", "(", "t", ".", "task_spec", ")", "]", ")", "while", "engine_steps", ":", "for", "task", "in", "engine_steps", ":", "task", ".", "complete", "(", ")", "engine_steps", "=", "list", "(", "[", "t", "for", "t", "in", "self", ".", "get_tasks", "(", "Task", ".", "READY", ")", "if", "self", ".", "_is_engine_task", "(", "t", ".", "task_spec", ")", "]", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BpmnWorkflow.refresh_waiting_tasks
Refresh the state of all WAITING tasks. This will, for example, update Catching Timer Events whose waiting time has passed.
SpiffWorkflow/bpmn/workflow.py
def refresh_waiting_tasks(self): """ Refresh the state of all WAITING tasks. This will, for example, update Catching Timer Events whose waiting time has passed. """ assert not self.read_only for my_task in self.get_tasks(Task.WAITING): my_task.task_spec._update(my_task)
def refresh_waiting_tasks(self): """ Refresh the state of all WAITING tasks. This will, for example, update Catching Timer Events whose waiting time has passed. """ assert not self.read_only for my_task in self.get_tasks(Task.WAITING): my_task.task_spec._update(my_task)
[ "Refresh", "the", "state", "of", "all", "WAITING", "tasks", ".", "This", "will", "for", "example", "update", "Catching", "Timer", "Events", "whose", "waiting", "time", "has", "passed", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/workflow.py#L81-L88
[ "def", "refresh_waiting_tasks", "(", "self", ")", ":", "assert", "not", "self", ".", "read_only", "for", "my_task", "in", "self", ".", "get_tasks", "(", "Task", ".", "WAITING", ")", ":", "my_task", ".", "task_spec", ".", "_update", "(", "my_task", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BpmnWorkflow.get_ready_user_tasks
Returns a list of User Tasks that are READY for user action
SpiffWorkflow/bpmn/workflow.py
def get_ready_user_tasks(self): """ Returns a list of User Tasks that are READY for user action """ return [t for t in self.get_tasks(Task.READY) if not self._is_engine_task(t.task_spec)]
def get_ready_user_tasks(self): """ Returns a list of User Tasks that are READY for user action """ return [t for t in self.get_tasks(Task.READY) if not self._is_engine_task(t.task_spec)]
[ "Returns", "a", "list", "of", "User", "Tasks", "that", "are", "READY", "for", "user", "action" ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/workflow.py#L90-L95
[ "def", "get_ready_user_tasks", "(", "self", ")", ":", "return", "[", "t", "for", "t", "in", "self", ".", "get_tasks", "(", "Task", ".", "READY", ")", "if", "not", "self", ".", "_is_engine_task", "(", "t", ".", "task_spec", ")", "]" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Trigger._on_trigger
Enqueue a trigger, such that this tasks triggers multiple times later when _on_complete() is called.
SpiffWorkflow/specs/Trigger.py
def _on_trigger(self, my_task): """ Enqueue a trigger, such that this tasks triggers multiple times later when _on_complete() is called. """ self.queued += 1 # All tasks that have already completed need to be put back to # READY. for thetask in my_task.workflow.task_tree: if thetask.thread_id != my_task.thread_id: continue if (thetask.task_spec == self and thetask._has_state(Task.COMPLETED)): thetask._set_state(Task.FUTURE, True) thetask._ready()
def _on_trigger(self, my_task): """ Enqueue a trigger, such that this tasks triggers multiple times later when _on_complete() is called. """ self.queued += 1 # All tasks that have already completed need to be put back to # READY. for thetask in my_task.workflow.task_tree: if thetask.thread_id != my_task.thread_id: continue if (thetask.task_spec == self and thetask._has_state(Task.COMPLETED)): thetask._set_state(Task.FUTURE, True) thetask._ready()
[ "Enqueue", "a", "trigger", "such", "that", "this", "tasks", "triggers", "multiple", "times", "later", "when", "_on_complete", "()", "is", "called", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/Trigger.py#L60-L74
[ "def", "_on_trigger", "(", "self", ",", "my_task", ")", ":", "self", ".", "queued", "+=", "1", "# All tasks that have already completed need to be put back to", "# READY.", "for", "thetask", "in", "my_task", ".", "workflow", ".", "task_tree", ":", "if", "thetask", ".", "thread_id", "!=", "my_task", ".", "thread_id", ":", "continue", "if", "(", "thetask", ".", "task_spec", "==", "self", "and", "thetask", ".", "_has_state", "(", "Task", ".", "COMPLETED", ")", ")", ":", "thetask", ".", "_set_state", "(", "Task", ".", "FUTURE", ",", "True", ")", "thetask", ".", "_ready", "(", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Trigger._on_complete_hook
A hook into _on_complete() that does the task specific work. :type my_task: Task :param my_task: A task in which this method is executed. :rtype: bool :returns: True on success, False otherwise.
SpiffWorkflow/specs/Trigger.py
def _on_complete_hook(self, my_task): """ A hook into _on_complete() that does the task specific work. :type my_task: Task :param my_task: A task in which this method is executed. :rtype: bool :returns: True on success, False otherwise. """ times = int(valueof(my_task, self.times, 1)) + self.queued for i in range(times): for task_name in self.context: task = my_task.workflow.get_task_spec_from_name(task_name) task._on_trigger(my_task) self.queued = 0 TaskSpec._on_complete_hook(self, my_task)
def _on_complete_hook(self, my_task): """ A hook into _on_complete() that does the task specific work. :type my_task: Task :param my_task: A task in which this method is executed. :rtype: bool :returns: True on success, False otherwise. """ times = int(valueof(my_task, self.times, 1)) + self.queued for i in range(times): for task_name in self.context: task = my_task.workflow.get_task_spec_from_name(task_name) task._on_trigger(my_task) self.queued = 0 TaskSpec._on_complete_hook(self, my_task)
[ "A", "hook", "into", "_on_complete", "()", "that", "does", "the", "task", "specific", "work", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/Trigger.py#L76-L91
[ "def", "_on_complete_hook", "(", "self", ",", "my_task", ")", ":", "times", "=", "int", "(", "valueof", "(", "my_task", ",", "self", ".", "times", ",", "1", ")", ")", "+", "self", ".", "queued", "for", "i", "in", "range", "(", "times", ")", ":", "for", "task_name", "in", "self", ".", "context", ":", "task", "=", "my_task", ".", "workflow", ".", "get_task_spec_from_name", "(", "task_name", ")", "task", ".", "_on_trigger", "(", "my_task", ")", "self", ".", "queued", "=", "0", "TaskSpec", ".", "_on_complete_hook", "(", "self", ",", "my_task", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Trigger.deserialize
Deserializes the trigger using the provided serializer.
SpiffWorkflow/specs/Trigger.py
def deserialize(cls, serializer, wf_spec, s_state, **kwargs): """ Deserializes the trigger using the provided serializer. """ return serializer.deserialize_trigger(wf_spec, s_state, **kwargs)
def deserialize(cls, serializer, wf_spec, s_state, **kwargs): """ Deserializes the trigger using the provided serializer. """ return serializer.deserialize_trigger(wf_spec, s_state, **kwargs)
[ "Deserializes", "the", "trigger", "using", "the", "provided", "serializer", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/Trigger.py#L97-L103
[ "def", "deserialize", "(", "cls", ",", "serializer", ",", "wf_spec", ",", "s_state", ",", "*", "*", "kwargs", ")", ":", "return", "serializer", ".", "deserialize_trigger", "(", "wf_spec", ",", "s_state", ",", "*", "*", "kwargs", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BpmnScriptEngine.evaluate
Evaluate the given expression, within the context of the given task and return the result.
SpiffWorkflow/bpmn/BpmnScriptEngine.py
def evaluate(self, task, expression): """ Evaluate the given expression, within the context of the given task and return the result. """ if isinstance(expression, Operator): return expression._matches(task) else: return self._eval(task, expression, **task.data)
def evaluate(self, task, expression): """ Evaluate the given expression, within the context of the given task and return the result. """ if isinstance(expression, Operator): return expression._matches(task) else: return self._eval(task, expression, **task.data)
[ "Evaluate", "the", "given", "expression", "within", "the", "context", "of", "the", "given", "task", "and", "return", "the", "result", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/BpmnScriptEngine.py#L37-L45
[ "def", "evaluate", "(", "self", ",", "task", ",", "expression", ")", ":", "if", "isinstance", "(", "expression", ",", "Operator", ")", ":", "return", "expression", ".", "_matches", "(", "task", ")", "else", ":", "return", "self", ".", "_eval", "(", "task", ",", "expression", ",", "*", "*", "task", ".", "data", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BpmnScriptEngine.execute
Execute the script, within the context of the specified task
SpiffWorkflow/bpmn/BpmnScriptEngine.py
def execute(self, task, script, **kwargs): """ Execute the script, within the context of the specified task """ locals().update(kwargs) exec(script)
def execute(self, task, script, **kwargs): """ Execute the script, within the context of the specified task """ locals().update(kwargs) exec(script)
[ "Execute", "the", "script", "within", "the", "context", "of", "the", "specified", "task" ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/bpmn/BpmnScriptEngine.py#L47-L52
[ "def", "execute", "(", "self", ",", "task", ",", "script", ",", "*", "*", "kwargs", ")", ":", "locals", "(", ")", ".", "update", "(", "kwargs", ")", "exec", "(", "script", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
merge_dictionary
Recursive merge two dicts (vs .update which overwrites the hashes at the root level) Note: This updates dst. Copied from checkmate.utils
SpiffWorkflow/util/__init__.py
def merge_dictionary(dst, src): """Recursive merge two dicts (vs .update which overwrites the hashes at the root level) Note: This updates dst. Copied from checkmate.utils """ stack = [(dst, src)] while stack: current_dst, current_src = stack.pop() for key in current_src: source = current_src[key] if key not in current_dst: current_dst[key] = source else: dest = current_dst[key] if isinstance(source, dict) and isinstance(dest, dict): stack.append((dest, source)) elif isinstance(source, list) and isinstance(dest, list): # Make them the same size r = dest[:] s = source[:] if len(dest) > len(source): s.append([None for i in range(len(dest) - len(source))]) elif len(dest) < len(source): r.append([None for i in range(len(source) - len(dest))]) # Merge lists for index, value in enumerate(r): if (not value) and s[index]: r[index] = s[index] elif isinstance(value, dict) and \ isinstance(s[index], dict): stack.append((dest[index], source[index])) else: dest[index] = s[index] current_dst[key] = r else: current_dst[key] = source return dst
def merge_dictionary(dst, src): """Recursive merge two dicts (vs .update which overwrites the hashes at the root level) Note: This updates dst. Copied from checkmate.utils """ stack = [(dst, src)] while stack: current_dst, current_src = stack.pop() for key in current_src: source = current_src[key] if key not in current_dst: current_dst[key] = source else: dest = current_dst[key] if isinstance(source, dict) and isinstance(dest, dict): stack.append((dest, source)) elif isinstance(source, list) and isinstance(dest, list): # Make them the same size r = dest[:] s = source[:] if len(dest) > len(source): s.append([None for i in range(len(dest) - len(source))]) elif len(dest) < len(source): r.append([None for i in range(len(source) - len(dest))]) # Merge lists for index, value in enumerate(r): if (not value) and s[index]: r[index] = s[index] elif isinstance(value, dict) and \ isinstance(s[index], dict): stack.append((dest[index], source[index])) else: dest[index] = s[index] current_dst[key] = r else: current_dst[key] = source return dst
[ "Recursive", "merge", "two", "dicts", "(", "vs", ".", "update", "which", "overwrites", "the", "hashes", "at", "the", "root", "level", ")" ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/util/__init__.py#L6-L46
[ "def", "merge_dictionary", "(", "dst", ",", "src", ")", ":", "stack", "=", "[", "(", "dst", ",", "src", ")", "]", "while", "stack", ":", "current_dst", ",", "current_src", "=", "stack", ".", "pop", "(", ")", "for", "key", "in", "current_src", ":", "source", "=", "current_src", "[", "key", "]", "if", "key", "not", "in", "current_dst", ":", "current_dst", "[", "key", "]", "=", "source", "else", ":", "dest", "=", "current_dst", "[", "key", "]", "if", "isinstance", "(", "source", ",", "dict", ")", "and", "isinstance", "(", "dest", ",", "dict", ")", ":", "stack", ".", "append", "(", "(", "dest", ",", "source", ")", ")", "elif", "isinstance", "(", "source", ",", "list", ")", "and", "isinstance", "(", "dest", ",", "list", ")", ":", "# Make them the same size", "r", "=", "dest", "[", ":", "]", "s", "=", "source", "[", ":", "]", "if", "len", "(", "dest", ")", ">", "len", "(", "source", ")", ":", "s", ".", "append", "(", "[", "None", "for", "i", "in", "range", "(", "len", "(", "dest", ")", "-", "len", "(", "source", ")", ")", "]", ")", "elif", "len", "(", "dest", ")", "<", "len", "(", "source", ")", ":", "r", ".", "append", "(", "[", "None", "for", "i", "in", "range", "(", "len", "(", "source", ")", "-", "len", "(", "dest", ")", ")", "]", ")", "# Merge lists", "for", "index", ",", "value", "in", "enumerate", "(", "r", ")", ":", "if", "(", "not", "value", ")", "and", "s", "[", "index", "]", ":", "r", "[", "index", "]", "=", "s", "[", "index", "]", "elif", "isinstance", "(", "value", ",", "dict", ")", "and", "isinstance", "(", "s", "[", "index", "]", ",", "dict", ")", ":", "stack", ".", "append", "(", "(", "dest", "[", "index", "]", ",", "source", "[", "index", "]", ")", ")", "else", ":", "dest", "[", "index", "]", "=", "s", "[", "index", "]", "current_dst", "[", "key", "]", "=", "r", "else", ":", "current_dst", "[", "key", "]", "=", "source", "return", "dst" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Join._start
Checks whether the preconditions for going to READY state are met. Returns True if the threshold was reached, False otherwise. Also returns the list of tasks that yet need to be completed.
SpiffWorkflow/specs/Join.py
def _start(self, my_task, force=False): """ Checks whether the preconditions for going to READY state are met. Returns True if the threshold was reached, False otherwise. Also returns the list of tasks that yet need to be completed. """ # If the threshold was already reached, there is nothing else to do. if my_task._has_state(Task.COMPLETED): return True, None if my_task._has_state(Task.READY): return True, None # Check whether we may fire. if self.split_task is None: return self._check_threshold_unstructured(my_task, force) return self._check_threshold_structured(my_task, force)
def _start(self, my_task, force=False): """ Checks whether the preconditions for going to READY state are met. Returns True if the threshold was reached, False otherwise. Also returns the list of tasks that yet need to be completed. """ # If the threshold was already reached, there is nothing else to do. if my_task._has_state(Task.COMPLETED): return True, None if my_task._has_state(Task.READY): return True, None # Check whether we may fire. if self.split_task is None: return self._check_threshold_unstructured(my_task, force) return self._check_threshold_structured(my_task, force)
[ "Checks", "whether", "the", "preconditions", "for", "going", "to", "READY", "state", "are", "met", ".", "Returns", "True", "if", "the", "threshold", "was", "reached", "False", "otherwise", ".", "Also", "returns", "the", "list", "of", "tasks", "that", "yet", "need", "to", "be", "completed", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/Join.py#L187-L202
[ "def", "_start", "(", "self", ",", "my_task", ",", "force", "=", "False", ")", ":", "# If the threshold was already reached, there is nothing else to do.", "if", "my_task", ".", "_has_state", "(", "Task", ".", "COMPLETED", ")", ":", "return", "True", ",", "None", "if", "my_task", ".", "_has_state", "(", "Task", ".", "READY", ")", ":", "return", "True", ",", "None", "# Check whether we may fire.", "if", "self", ".", "split_task", "is", "None", ":", "return", "self", ".", "_check_threshold_unstructured", "(", "my_task", ",", "force", ")", "return", "self", ".", "_check_threshold_structured", "(", "my_task", ",", "force", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
Join._on_trigger
May be called to fire the Join before the incoming branches are completed.
SpiffWorkflow/specs/Join.py
def _on_trigger(self, my_task): """ May be called to fire the Join before the incoming branches are completed. """ for task in my_task.workflow.task_tree._find_any(self): if task.thread_id != my_task.thread_id: continue self._do_join(task)
def _on_trigger(self, my_task): """ May be called to fire the Join before the incoming branches are completed. """ for task in my_task.workflow.task_tree._find_any(self): if task.thread_id != my_task.thread_id: continue self._do_join(task)
[ "May", "be", "called", "to", "fire", "the", "Join", "before", "the", "incoming", "branches", "are", "completed", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/Join.py#L284-L292
[ "def", "_on_trigger", "(", "self", ",", "my_task", ")", ":", "for", "task", "in", "my_task", ".", "workflow", ".", "task_tree", ".", "_find_any", "(", "self", ")", ":", "if", "task", ".", "thread_id", "!=", "my_task", ".", "thread_id", ":", "continue", "self", ".", "_do_join", "(", "task", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
ExclusiveChoice.connect
Connects the task spec that is executed if no other condition matches. :type task_spec: TaskSpec :param task_spec: The following task spec.
SpiffWorkflow/specs/ExclusiveChoice.py
def connect(self, task_spec): """ Connects the task spec that is executed if no other condition matches. :type task_spec: TaskSpec :param task_spec: The following task spec. """ assert self.default_task_spec is None self.outputs.append(task_spec) self.default_task_spec = task_spec.name task_spec._connect_notify(self)
def connect(self, task_spec): """ Connects the task spec that is executed if no other condition matches. :type task_spec: TaskSpec :param task_spec: The following task spec. """ assert self.default_task_spec is None self.outputs.append(task_spec) self.default_task_spec = task_spec.name task_spec._connect_notify(self)
[ "Connects", "the", "task", "spec", "that", "is", "executed", "if", "no", "other", "condition", "matches", "." ]
knipknap/SpiffWorkflow
python
https://github.com/knipknap/SpiffWorkflow/blob/f0af7f59a332e0619e4f3c00a7d4a3d230760e00/SpiffWorkflow/specs/ExclusiveChoice.py#L47-L58
[ "def", "connect", "(", "self", ",", "task_spec", ")", ":", "assert", "self", ".", "default_task_spec", "is", "None", "self", ".", "outputs", ".", "append", "(", "task_spec", ")", "self", ".", "default_task_spec", "=", "task_spec", ".", "name", "task_spec", ".", "_connect_notify", "(", "self", ")" ]
f0af7f59a332e0619e4f3c00a7d4a3d230760e00
valid
BlockadeState.container_id
Try to find the container ID with the specified name
blockade/state.py
def container_id(self, name): '''Try to find the container ID with the specified name''' container = self._containers.get(name, None) if not container is None: return container.get('id', None) return None
def container_id(self, name): '''Try to find the container ID with the specified name''' container = self._containers.get(name, None) if not container is None: return container.get('id', None) return None
[ "Try", "to", "find", "the", "container", "ID", "with", "the", "specified", "name" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L84-L89
[ "def", "container_id", "(", "self", ",", "name", ")", ":", "container", "=", "self", ".", "_containers", ".", "get", "(", "name", ",", "None", ")", "if", "not", "container", "is", "None", ":", "return", "container", ".", "get", "(", "'id'", ",", "None", ")", "return", "None" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState.initialize
Initialize a new state file with the given contents. This function fails in case the state file already exists.
blockade/state.py
def initialize(self, containers): ''' Initialize a new state file with the given contents. This function fails in case the state file already exists. ''' self._containers = deepcopy(containers) self.__write(containers, initialize=True)
def initialize(self, containers): ''' Initialize a new state file with the given contents. This function fails in case the state file already exists. ''' self._containers = deepcopy(containers) self.__write(containers, initialize=True)
[ "Initialize", "a", "new", "state", "file", "with", "the", "given", "contents", ".", "This", "function", "fails", "in", "case", "the", "state", "file", "already", "exists", "." ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L91-L97
[ "def", "initialize", "(", "self", ",", "containers", ")", ":", "self", ".", "_containers", "=", "deepcopy", "(", "containers", ")", "self", ".", "__write", "(", "containers", ",", "initialize", "=", "True", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState.update
Update the current state file with the specified contents
blockade/state.py
def update(self, containers): '''Update the current state file with the specified contents''' self._containers = deepcopy(containers) self.__write(containers, initialize=False)
def update(self, containers): '''Update the current state file with the specified contents''' self._containers = deepcopy(containers) self.__write(containers, initialize=False)
[ "Update", "the", "current", "state", "file", "with", "the", "specified", "contents" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L103-L106
[ "def", "update", "(", "self", ",", "containers", ")", ":", "self", ".", "_containers", "=", "deepcopy", "(", "containers", ")", "self", ".", "__write", "(", "containers", ",", "initialize", "=", "False", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState.load
Try to load a blockade state file in the current directory
blockade/state.py
def load(self): '''Try to load a blockade state file in the current directory''' try: with open(self._state_file) as f: state = yaml.safe_load(f) self._containers = state['containers'] except (IOError, OSError) as err: if err.errno == errno.ENOENT: raise NotInitializedError("No blockade exists in this context") raise InconsistentStateError("Failed to load Blockade state: " + str(err)) except Exception as err: raise InconsistentStateError("Failed to load Blockade state: " + str(err))
def load(self): '''Try to load a blockade state file in the current directory''' try: with open(self._state_file) as f: state = yaml.safe_load(f) self._containers = state['containers'] except (IOError, OSError) as err: if err.errno == errno.ENOENT: raise NotInitializedError("No blockade exists in this context") raise InconsistentStateError("Failed to load Blockade state: " + str(err)) except Exception as err: raise InconsistentStateError("Failed to load Blockade state: " + str(err))
[ "Try", "to", "load", "a", "blockade", "state", "file", "in", "the", "current", "directory" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L108-L121
[ "def", "load", "(", "self", ")", ":", "try", ":", "with", "open", "(", "self", ".", "_state_file", ")", "as", "f", ":", "state", "=", "yaml", ".", "safe_load", "(", "f", ")", "self", ".", "_containers", "=", "state", "[", "'containers'", "]", "except", "(", "IOError", ",", "OSError", ")", "as", "err", ":", "if", "err", ".", "errno", "==", "errno", ".", "ENOENT", ":", "raise", "NotInitializedError", "(", "\"No blockade exists in this context\"", ")", "raise", "InconsistentStateError", "(", "\"Failed to load Blockade state: \"", "+", "str", "(", "err", ")", ")", "except", "Exception", "as", "err", ":", "raise", "InconsistentStateError", "(", "\"Failed to load Blockade state: \"", "+", "str", "(", "err", ")", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState._get_blockade_id_from_cwd
Generate a new blockade ID based on the CWD
blockade/state.py
def _get_blockade_id_from_cwd(self, cwd=None): '''Generate a new blockade ID based on the CWD''' if not cwd: cwd = os.getcwd() # this follows a similar pattern as docker-compose uses parent_dir = os.path.abspath(cwd) basename = os.path.basename(parent_dir).lower() blockade_id = re.sub(r"[^a-z0-9]", "", basename) if not blockade_id: # if we can't get a valid name from CWD, use "default" blockade_id = "default" return blockade_id
def _get_blockade_id_from_cwd(self, cwd=None): '''Generate a new blockade ID based on the CWD''' if not cwd: cwd = os.getcwd() # this follows a similar pattern as docker-compose uses parent_dir = os.path.abspath(cwd) basename = os.path.basename(parent_dir).lower() blockade_id = re.sub(r"[^a-z0-9]", "", basename) if not blockade_id: # if we can't get a valid name from CWD, use "default" blockade_id = "default" return blockade_id
[ "Generate", "a", "new", "blockade", "ID", "based", "on", "the", "CWD" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L127-L137
[ "def", "_get_blockade_id_from_cwd", "(", "self", ",", "cwd", "=", "None", ")", ":", "if", "not", "cwd", ":", "cwd", "=", "os", ".", "getcwd", "(", ")", "# this follows a similar pattern as docker-compose uses", "parent_dir", "=", "os", ".", "path", ".", "abspath", "(", "cwd", ")", "basename", "=", "os", ".", "path", ".", "basename", "(", "parent_dir", ")", ".", "lower", "(", ")", "blockade_id", "=", "re", ".", "sub", "(", "r\"[^a-z0-9]\"", ",", "\"\"", ",", "basename", ")", "if", "not", "blockade_id", ":", "# if we can't get a valid name from CWD, use \"default\"", "blockade_id", "=", "\"default\"", "return", "blockade_id" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState._assure_dir
Make sure the state directory exists
blockade/state.py
def _assure_dir(self): '''Make sure the state directory exists''' try: os.makedirs(self._state_dir) except OSError as err: if err.errno != errno.EEXIST: raise
def _assure_dir(self): '''Make sure the state directory exists''' try: os.makedirs(self._state_dir) except OSError as err: if err.errno != errno.EEXIST: raise
[ "Make", "sure", "the", "state", "directory", "exists" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L139-L145
[ "def", "_assure_dir", "(", "self", ")", ":", "try", ":", "os", ".", "makedirs", "(", "self", ".", "_state_dir", ")", "except", "OSError", "as", "err", ":", "if", "err", ".", "errno", "!=", "errno", ".", "EEXIST", ":", "raise" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState._state_delete
Try to delete the state.yml file and the folder .blockade
blockade/state.py
def _state_delete(self): '''Try to delete the state.yml file and the folder .blockade''' try: os.remove(self._state_file) except OSError as err: if err.errno not in (errno.EPERM, errno.ENOENT): raise try: os.rmdir(self._state_dir) except OSError as err: if err.errno not in (errno.ENOTEMPTY, errno.ENOENT): raise
def _state_delete(self): '''Try to delete the state.yml file and the folder .blockade''' try: os.remove(self._state_file) except OSError as err: if err.errno not in (errno.EPERM, errno.ENOENT): raise try: os.rmdir(self._state_dir) except OSError as err: if err.errno not in (errno.ENOTEMPTY, errno.ENOENT): raise
[ "Try", "to", "delete", "the", "state", ".", "yml", "file", "and", "the", "folder", ".", "blockade" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L147-L159
[ "def", "_state_delete", "(", "self", ")", ":", "try", ":", "os", ".", "remove", "(", "self", ".", "_state_file", ")", "except", "OSError", "as", "err", ":", "if", "err", ".", "errno", "not", "in", "(", "errno", ".", "EPERM", ",", "errno", ".", "ENOENT", ")", ":", "raise", "try", ":", "os", ".", "rmdir", "(", "self", ".", "_state_dir", ")", "except", "OSError", "as", "err", ":", "if", "err", ".", "errno", "not", "in", "(", "errno", ".", "ENOTEMPTY", ",", "errno", ".", "ENOENT", ")", ":", "raise" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState.__base_state
Convert blockade ID and container information into a state dictionary object.
blockade/state.py
def __base_state(self, containers): ''' Convert blockade ID and container information into a state dictionary object. ''' return dict(blockade_id=self._blockade_id, containers=containers, version=self._state_version)
def __base_state(self, containers): ''' Convert blockade ID and container information into a state dictionary object. ''' return dict(blockade_id=self._blockade_id, containers=containers, version=self._state_version)
[ "Convert", "blockade", "ID", "and", "container", "information", "into", "a", "state", "dictionary", "object", "." ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L161-L168
[ "def", "__base_state", "(", "self", ",", "containers", ")", ":", "return", "dict", "(", "blockade_id", "=", "self", ".", "_blockade_id", ",", "containers", "=", "containers", ",", "version", "=", "self", ".", "_state_version", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeState.__write
Write the given state information into a file
blockade/state.py
def __write(self, containers, initialize=True): '''Write the given state information into a file''' path = self._state_file self._assure_dir() try: flags = os.O_WRONLY | os.O_CREAT if initialize: flags |= os.O_EXCL with os.fdopen(os.open(path, flags), "w") as f: yaml.safe_dump(self.__base_state(containers), f) except OSError as err: if err.errno == errno.EEXIST: raise AlreadyInitializedError( "Path %s exists. " "You may need to destroy a previous blockade." % path) raise except Exception: # clean up our created file self._state_delete() raise
def __write(self, containers, initialize=True): '''Write the given state information into a file''' path = self._state_file self._assure_dir() try: flags = os.O_WRONLY | os.O_CREAT if initialize: flags |= os.O_EXCL with os.fdopen(os.open(path, flags), "w") as f: yaml.safe_dump(self.__base_state(containers), f) except OSError as err: if err.errno == errno.EEXIST: raise AlreadyInitializedError( "Path %s exists. " "You may need to destroy a previous blockade." % path) raise except Exception: # clean up our created file self._state_delete() raise
[ "Write", "the", "given", "state", "information", "into", "a", "file" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/state.py#L170-L189
[ "def", "__write", "(", "self", ",", "containers", ",", "initialize", "=", "True", ")", ":", "path", "=", "self", ".", "_state_file", "self", ".", "_assure_dir", "(", ")", "try", ":", "flags", "=", "os", ".", "O_WRONLY", "|", "os", ".", "O_CREAT", "if", "initialize", ":", "flags", "|=", "os", ".", "O_EXCL", "with", "os", ".", "fdopen", "(", "os", ".", "open", "(", "path", ",", "flags", ")", ",", "\"w\"", ")", "as", "f", ":", "yaml", ".", "safe_dump", "(", "self", ".", "__base_state", "(", "containers", ")", ",", "f", ")", "except", "OSError", "as", "err", ":", "if", "err", ".", "errno", "==", "errno", ".", "EEXIST", ":", "raise", "AlreadyInitializedError", "(", "\"Path %s exists. \"", "\"You may need to destroy a previous blockade.\"", "%", "path", ")", "raise", "except", "Exception", ":", "# clean up our created file", "self", ".", "_state_delete", "(", ")", "raise" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
expand_partitions
Validate the partitions of containers. If there are any containers not in any partition, place them in an new partition.
blockade/core.py
def expand_partitions(containers, partitions): ''' Validate the partitions of containers. If there are any containers not in any partition, place them in an new partition. ''' # filter out holy containers that don't belong # to any partition at all all_names = frozenset(c.name for c in containers if not c.holy) holy_names = frozenset(c.name for c in containers if c.holy) neutral_names = frozenset(c.name for c in containers if c.neutral) partitions = [frozenset(p) for p in partitions] unknown = set() holy = set() union = set() for partition in partitions: unknown.update(partition - all_names - holy_names) holy.update(partition - all_names) union.update(partition) if unknown: raise BlockadeError('Partitions contain unknown containers: %s' % list(unknown)) if holy: raise BlockadeError('Partitions contain holy containers: %s' % list(holy)) # put any leftover containers in an implicit partition leftover = all_names.difference(union) if leftover: partitions.append(leftover) # we create an 'implicit' partition for the neutral containers # in case they are not part of the leftover anyways if not neutral_names.issubset(leftover): partitions.append(neutral_names) return partitions
def expand_partitions(containers, partitions): ''' Validate the partitions of containers. If there are any containers not in any partition, place them in an new partition. ''' # filter out holy containers that don't belong # to any partition at all all_names = frozenset(c.name for c in containers if not c.holy) holy_names = frozenset(c.name for c in containers if c.holy) neutral_names = frozenset(c.name for c in containers if c.neutral) partitions = [frozenset(p) for p in partitions] unknown = set() holy = set() union = set() for partition in partitions: unknown.update(partition - all_names - holy_names) holy.update(partition - all_names) union.update(partition) if unknown: raise BlockadeError('Partitions contain unknown containers: %s' % list(unknown)) if holy: raise BlockadeError('Partitions contain holy containers: %s' % list(holy)) # put any leftover containers in an implicit partition leftover = all_names.difference(union) if leftover: partitions.append(leftover) # we create an 'implicit' partition for the neutral containers # in case they are not part of the leftover anyways if not neutral_names.issubset(leftover): partitions.append(neutral_names) return partitions
[ "Validate", "the", "partitions", "of", "containers", ".", "If", "there", "are", "any", "containers", "not", "in", "any", "partition", "place", "them", "in", "an", "new", "partition", "." ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/core.py#L600-L640
[ "def", "expand_partitions", "(", "containers", ",", "partitions", ")", ":", "# filter out holy containers that don't belong", "# to any partition at all", "all_names", "=", "frozenset", "(", "c", ".", "name", "for", "c", "in", "containers", "if", "not", "c", ".", "holy", ")", "holy_names", "=", "frozenset", "(", "c", ".", "name", "for", "c", "in", "containers", "if", "c", ".", "holy", ")", "neutral_names", "=", "frozenset", "(", "c", ".", "name", "for", "c", "in", "containers", "if", "c", ".", "neutral", ")", "partitions", "=", "[", "frozenset", "(", "p", ")", "for", "p", "in", "partitions", "]", "unknown", "=", "set", "(", ")", "holy", "=", "set", "(", ")", "union", "=", "set", "(", ")", "for", "partition", "in", "partitions", ":", "unknown", ".", "update", "(", "partition", "-", "all_names", "-", "holy_names", ")", "holy", ".", "update", "(", "partition", "-", "all_names", ")", "union", ".", "update", "(", "partition", ")", "if", "unknown", ":", "raise", "BlockadeError", "(", "'Partitions contain unknown containers: %s'", "%", "list", "(", "unknown", ")", ")", "if", "holy", ":", "raise", "BlockadeError", "(", "'Partitions contain holy containers: %s'", "%", "list", "(", "holy", ")", ")", "# put any leftover containers in an implicit partition", "leftover", "=", "all_names", ".", "difference", "(", "union", ")", "if", "leftover", ":", "partitions", ".", "append", "(", "leftover", ")", "# we create an 'implicit' partition for the neutral containers", "# in case they are not part of the leftover anyways", "if", "not", "neutral_names", ".", "issubset", "(", "leftover", ")", ":", "partitions", ".", "append", "(", "neutral_names", ")", "return", "partitions" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
_IPTables.get_source_chains
Get a map of blockade chains IDs -> list of IPs targeted at them For figuring out which container is in which partition
blockade/net.py
def get_source_chains(self, blockade_id): """Get a map of blockade chains IDs -> list of IPs targeted at them For figuring out which container is in which partition """ result = {} if not blockade_id: raise ValueError("invalid blockade_id") lines = self.get_chain_rules("FORWARD") for line in lines: parts = line.split() if len(parts) < 4: continue try: partition_index = parse_partition_index(blockade_id, parts[0]) except ValueError: continue # not a rule targetting a blockade chain source = parts[3] if source: result[source] = partition_index return result
def get_source_chains(self, blockade_id): """Get a map of blockade chains IDs -> list of IPs targeted at them For figuring out which container is in which partition """ result = {} if not blockade_id: raise ValueError("invalid blockade_id") lines = self.get_chain_rules("FORWARD") for line in lines: parts = line.split() if len(parts) < 4: continue try: partition_index = parse_partition_index(blockade_id, parts[0]) except ValueError: continue # not a rule targetting a blockade chain source = parts[3] if source: result[source] = partition_index return result
[ "Get", "a", "map", "of", "blockade", "chains", "IDs", "-", ">", "list", "of", "IPs", "targeted", "at", "them" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/net.py#L169-L191
[ "def", "get_source_chains", "(", "self", ",", "blockade_id", ")", ":", "result", "=", "{", "}", "if", "not", "blockade_id", ":", "raise", "ValueError", "(", "\"invalid blockade_id\"", ")", "lines", "=", "self", ".", "get_chain_rules", "(", "\"FORWARD\"", ")", "for", "line", "in", "lines", ":", "parts", "=", "line", ".", "split", "(", ")", "if", "len", "(", "parts", ")", "<", "4", ":", "continue", "try", ":", "partition_index", "=", "parse_partition_index", "(", "blockade_id", ",", "parts", "[", "0", "]", ")", "except", "ValueError", ":", "continue", "# not a rule targetting a blockade chain", "source", "=", "parts", "[", "3", "]", "if", "source", ":", "result", "[", "source", "]", "=", "partition_index", "return", "result" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
_IPTables.insert_rule
Insert a new rule in the chain
blockade/net.py
def insert_rule(self, chain, src=None, dest=None, target=None): """Insert a new rule in the chain """ if not chain: raise ValueError("Invalid chain") if not target: raise ValueError("Invalid target") if not (src or dest): raise ValueError("Need src, dest, or both") args = ["-I", chain] if src: args += ["-s", src] if dest: args += ["-d", dest] args += ["-j", target] self.call(*args)
def insert_rule(self, chain, src=None, dest=None, target=None): """Insert a new rule in the chain """ if not chain: raise ValueError("Invalid chain") if not target: raise ValueError("Invalid target") if not (src or dest): raise ValueError("Need src, dest, or both") args = ["-I", chain] if src: args += ["-s", src] if dest: args += ["-d", dest] args += ["-j", target] self.call(*args)
[ "Insert", "a", "new", "rule", "in", "the", "chain" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/net.py#L235-L251
[ "def", "insert_rule", "(", "self", ",", "chain", ",", "src", "=", "None", ",", "dest", "=", "None", ",", "target", "=", "None", ")", ":", "if", "not", "chain", ":", "raise", "ValueError", "(", "\"Invalid chain\"", ")", "if", "not", "target", ":", "raise", "ValueError", "(", "\"Invalid target\"", ")", "if", "not", "(", "src", "or", "dest", ")", ":", "raise", "ValueError", "(", "\"Need src, dest, or both\"", ")", "args", "=", "[", "\"-I\"", ",", "chain", "]", "if", "src", ":", "args", "+=", "[", "\"-s\"", ",", "src", "]", "if", "dest", ":", "args", "+=", "[", "\"-d\"", ",", "dest", "]", "args", "+=", "[", "\"-j\"", ",", "target", "]", "self", ".", "call", "(", "*", "args", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeChaos._sm_start
Start the timer waiting for pain
blockade/chaos.py
def _sm_start(self, *args, **kwargs): """ Start the timer waiting for pain """ millisec = random.randint(self._start_min_delay, self._start_max_delay) self._timer = threading.Timer(millisec / 1000.0, self.event_timeout) self._timer.start()
def _sm_start(self, *args, **kwargs): """ Start the timer waiting for pain """ millisec = random.randint(self._start_min_delay, self._start_max_delay) self._timer = threading.Timer(millisec / 1000.0, self.event_timeout) self._timer.start()
[ "Start", "the", "timer", "waiting", "for", "pain" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/chaos.py#L286-L292
[ "def", "_sm_start", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "millisec", "=", "random", ".", "randint", "(", "self", ".", "_start_min_delay", ",", "self", ".", "_start_max_delay", ")", "self", ".", "_timer", "=", "threading", ".", "Timer", "(", "millisec", "/", "1000.0", ",", "self", ".", "event_timeout", ")", "self", ".", "_timer", ".", "start", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeChaos._sm_to_pain
Start the blockade event
blockade/chaos.py
def _sm_to_pain(self, *args, **kwargs): """ Start the blockade event """ _logger.info("Starting chaos for blockade %s" % self._blockade_name) self._do_blockade_event() # start the timer to end the pain millisec = random.randint(self._run_min_time, self._run_max_time) self._timer = threading.Timer(millisec / 1000.0, self.event_timeout) self._timer.start()
def _sm_to_pain(self, *args, **kwargs): """ Start the blockade event """ _logger.info("Starting chaos for blockade %s" % self._blockade_name) self._do_blockade_event() # start the timer to end the pain millisec = random.randint(self._run_min_time, self._run_max_time) self._timer = threading.Timer(millisec / 1000.0, self.event_timeout) self._timer.start()
[ "Start", "the", "blockade", "event" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/chaos.py#L294-L303
[ "def", "_sm_to_pain", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "_logger", ".", "info", "(", "\"Starting chaos for blockade %s\"", "%", "self", ".", "_blockade_name", ")", "self", ".", "_do_blockade_event", "(", ")", "# start the timer to end the pain", "millisec", "=", "random", ".", "randint", "(", "self", ".", "_run_min_time", ",", "self", ".", "_run_max_time", ")", "self", ".", "_timer", "=", "threading", ".", "Timer", "(", "millisec", "/", "1000.0", ",", "self", ".", "event_timeout", ")", "self", ".", "_timer", ".", "start", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeChaos._sm_stop_from_no_pain
Stop chaos when there is no current blockade operation
blockade/chaos.py
def _sm_stop_from_no_pain(self, *args, **kwargs): """ Stop chaos when there is no current blockade operation """ # Just stop the timer. It is possible that it was too late and the # timer is about to run _logger.info("Stopping chaos for blockade %s" % self._blockade_name) self._timer.cancel()
def _sm_stop_from_no_pain(self, *args, **kwargs): """ Stop chaos when there is no current blockade operation """ # Just stop the timer. It is possible that it was too late and the # timer is about to run _logger.info("Stopping chaos for blockade %s" % self._blockade_name) self._timer.cancel()
[ "Stop", "chaos", "when", "there", "is", "no", "current", "blockade", "operation" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/chaos.py#L305-L312
[ "def", "_sm_stop_from_no_pain", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "# Just stop the timer. It is possible that it was too late and the", "# timer is about to run", "_logger", ".", "info", "(", "\"Stopping chaos for blockade %s\"", "%", "self", ".", "_blockade_name", ")", "self", ".", "_timer", ".", "cancel", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeChaos._sm_relieve_pain
End the blockade event and return to a steady state
blockade/chaos.py
def _sm_relieve_pain(self, *args, **kwargs): """ End the blockade event and return to a steady state """ _logger.info( "Ending the degradation for blockade %s" % self._blockade_name) self._do_reset_all() # set a timer for the next pain event millisec = random.randint(self._start_min_delay, self._start_max_delay) self._timer = threading.Timer(millisec/1000.0, self.event_timeout) self._timer.start()
def _sm_relieve_pain(self, *args, **kwargs): """ End the blockade event and return to a steady state """ _logger.info( "Ending the degradation for blockade %s" % self._blockade_name) self._do_reset_all() # set a timer for the next pain event millisec = random.randint(self._start_min_delay, self._start_max_delay) self._timer = threading.Timer(millisec/1000.0, self.event_timeout) self._timer.start()
[ "End", "the", "blockade", "event", "and", "return", "to", "a", "steady", "state" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/chaos.py#L314-L324
[ "def", "_sm_relieve_pain", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "_logger", ".", "info", "(", "\"Ending the degradation for blockade %s\"", "%", "self", ".", "_blockade_name", ")", "self", ".", "_do_reset_all", "(", ")", "# set a timer for the next pain event", "millisec", "=", "random", ".", "randint", "(", "self", ".", "_start_min_delay", ",", "self", ".", "_start_max_delay", ")", "self", ".", "_timer", "=", "threading", ".", "Timer", "(", "millisec", "/", "1000.0", ",", "self", ".", "event_timeout", ")", "self", ".", "_timer", ".", "start", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeChaos._sm_stop_from_pain
Stop chaos while there is a blockade event in progress
blockade/chaos.py
def _sm_stop_from_pain(self, *args, **kwargs): """ Stop chaos while there is a blockade event in progress """ _logger.info("Stopping chaos for blockade %s" % self._blockade_name) self._do_reset_all()
def _sm_stop_from_pain(self, *args, **kwargs): """ Stop chaos while there is a blockade event in progress """ _logger.info("Stopping chaos for blockade %s" % self._blockade_name) self._do_reset_all()
[ "Stop", "chaos", "while", "there", "is", "a", "blockade", "event", "in", "progress" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/chaos.py#L326-L331
[ "def", "_sm_stop_from_pain", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "_logger", ".", "info", "(", "\"Stopping chaos for blockade %s\"", "%", "self", ".", "_blockade_name", ")", "self", ".", "_do_reset_all", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeChaos._sm_cleanup
Delete all state associated with the chaos session
blockade/chaos.py
def _sm_cleanup(self, *args, **kwargs): """ Delete all state associated with the chaos session """ if self._done_notification_func is not None: self._done_notification_func() self._timer.cancel()
def _sm_cleanup(self, *args, **kwargs): """ Delete all state associated with the chaos session """ if self._done_notification_func is not None: self._done_notification_func() self._timer.cancel()
[ "Delete", "all", "state", "associated", "with", "the", "chaos", "session" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/chaos.py#L333-L339
[ "def", "_sm_cleanup", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "self", ".", "_done_notification_func", "is", "not", "None", ":", "self", ".", "_done_notification_func", "(", ")", "self", ".", "_timer", ".", "cancel", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
dependency_sorted
Sort a dictionary or list of containers into dependency order Returns a sequence
blockade/config.py
def dependency_sorted(containers): """Sort a dictionary or list of containers into dependency order Returns a sequence """ if not isinstance(containers, collections.Mapping): containers = dict((c.name, c) for c in containers) container_links = dict((name, set(c.links.keys())) for name, c in containers.items()) sorted_names = _resolve(container_links) return [containers[name] for name in sorted_names]
def dependency_sorted(containers): """Sort a dictionary or list of containers into dependency order Returns a sequence """ if not isinstance(containers, collections.Mapping): containers = dict((c.name, c) for c in containers) container_links = dict((name, set(c.links.keys())) for name, c in containers.items()) sorted_names = _resolve(container_links) return [containers[name] for name in sorted_names]
[ "Sort", "a", "dictionary", "or", "list", "of", "containers", "into", "dependency", "order" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/config.py#L206-L217
[ "def", "dependency_sorted", "(", "containers", ")", ":", "if", "not", "isinstance", "(", "containers", ",", "collections", ".", "Mapping", ")", ":", "containers", "=", "dict", "(", "(", "c", ".", "name", ",", "c", ")", "for", "c", "in", "containers", ")", "container_links", "=", "dict", "(", "(", "name", ",", "set", "(", "c", ".", "links", ".", "keys", "(", ")", ")", ")", "for", "name", ",", "c", "in", "containers", ".", "items", "(", ")", ")", "sorted_names", "=", "_resolve", "(", "container_links", ")", "return", "[", "containers", "[", "name", "]", "for", "name", "in", "sorted_names", "]" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeContainerConfig.from_dict
Convert a dictionary of configuration values into a sequence of BlockadeContainerConfig instances
blockade/config.py
def from_dict(name, values): ''' Convert a dictionary of configuration values into a sequence of BlockadeContainerConfig instances ''' # determine the number of instances of this container count = 1 count_value = values.get('count', 1) if isinstance(count_value, int): count = max(count_value, 1) def with_index(name, idx): if name and idx: return '%s_%d' % (name, idx) return name def get_instance(n, idx=None): return BlockadeContainerConfig( with_index(n, idx), values['image'], command=values.get('command'), links=values.get('links'), volumes=values.get('volumes'), publish_ports=values.get('ports'), expose_ports=values.get('expose'), environment=values.get('environment'), hostname=values.get('hostname'), dns=values.get('dns'), start_delay=values.get('start_delay', 0), neutral=values.get('neutral', False), holy=values.get('holy', False), container_name=with_index(values.get('container_name'), idx), cap_add=values.get('cap_add')) if count == 1: yield get_instance(name) else: for idx in range(1, count+1): # TODO: configurable name/index format yield get_instance(name, idx)
def from_dict(name, values): ''' Convert a dictionary of configuration values into a sequence of BlockadeContainerConfig instances ''' # determine the number of instances of this container count = 1 count_value = values.get('count', 1) if isinstance(count_value, int): count = max(count_value, 1) def with_index(name, idx): if name and idx: return '%s_%d' % (name, idx) return name def get_instance(n, idx=None): return BlockadeContainerConfig( with_index(n, idx), values['image'], command=values.get('command'), links=values.get('links'), volumes=values.get('volumes'), publish_ports=values.get('ports'), expose_ports=values.get('expose'), environment=values.get('environment'), hostname=values.get('hostname'), dns=values.get('dns'), start_delay=values.get('start_delay', 0), neutral=values.get('neutral', False), holy=values.get('holy', False), container_name=with_index(values.get('container_name'), idx), cap_add=values.get('cap_add')) if count == 1: yield get_instance(name) else: for idx in range(1, count+1): # TODO: configurable name/index format yield get_instance(name, idx)
[ "Convert", "a", "dictionary", "of", "configuration", "values", "into", "a", "sequence", "of", "BlockadeContainerConfig", "instances" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/config.py#L30-L70
[ "def", "from_dict", "(", "name", ",", "values", ")", ":", "# determine the number of instances of this container", "count", "=", "1", "count_value", "=", "values", ".", "get", "(", "'count'", ",", "1", ")", "if", "isinstance", "(", "count_value", ",", "int", ")", ":", "count", "=", "max", "(", "count_value", ",", "1", ")", "def", "with_index", "(", "name", ",", "idx", ")", ":", "if", "name", "and", "idx", ":", "return", "'%s_%d'", "%", "(", "name", ",", "idx", ")", "return", "name", "def", "get_instance", "(", "n", ",", "idx", "=", "None", ")", ":", "return", "BlockadeContainerConfig", "(", "with_index", "(", "n", ",", "idx", ")", ",", "values", "[", "'image'", "]", ",", "command", "=", "values", ".", "get", "(", "'command'", ")", ",", "links", "=", "values", ".", "get", "(", "'links'", ")", ",", "volumes", "=", "values", ".", "get", "(", "'volumes'", ")", ",", "publish_ports", "=", "values", ".", "get", "(", "'ports'", ")", ",", "expose_ports", "=", "values", ".", "get", "(", "'expose'", ")", ",", "environment", "=", "values", ".", "get", "(", "'environment'", ")", ",", "hostname", "=", "values", ".", "get", "(", "'hostname'", ")", ",", "dns", "=", "values", ".", "get", "(", "'dns'", ")", ",", "start_delay", "=", "values", ".", "get", "(", "'start_delay'", ",", "0", ")", ",", "neutral", "=", "values", ".", "get", "(", "'neutral'", ",", "False", ")", ",", "holy", "=", "values", ".", "get", "(", "'holy'", ",", "False", ")", ",", "container_name", "=", "with_index", "(", "values", ".", "get", "(", "'container_name'", ")", ",", "idx", ")", ",", "cap_add", "=", "values", ".", "get", "(", "'cap_add'", ")", ")", "if", "count", "==", "1", ":", "yield", "get_instance", "(", "name", ")", "else", ":", "for", "idx", "in", "range", "(", "1", ",", "count", "+", "1", ")", ":", "# TODO: configurable name/index format", "yield", "get_instance", "(", "name", ",", "idx", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
BlockadeConfig.from_dict
Instantiate a BlockadeConfig instance based on a given dictionary of configuration values
blockade/config.py
def from_dict(values): ''' Instantiate a BlockadeConfig instance based on a given dictionary of configuration values ''' try: containers = values['containers'] parsed_containers = {} for name, container_dict in containers.items(): try: # one config entry might result in many container # instances (indicated by the 'count' config value) for cnt in BlockadeContainerConfig.from_dict(name, container_dict): # check for duplicate 'container_name' definitions if cnt.container_name: cname = cnt.container_name existing = [c for c in parsed_containers.values() if c.container_name == cname] if existing: raise BlockadeConfigError("Duplicate 'container_name' definition: %s" % (cname)) parsed_containers[cnt.name] = cnt except Exception as err: raise BlockadeConfigError( "Container '%s' config problem: %s" % (name, err)) network = values.get('network') if network: defaults = _DEFAULT_NETWORK_CONFIG.copy() defaults.update(network) network = defaults else: network = _DEFAULT_NETWORK_CONFIG.copy() return BlockadeConfig(parsed_containers, network=network) except KeyError as err: raise BlockadeConfigError("Config missing value: " + str(err)) except Exception as err: # TODO log this to some debug stream? raise BlockadeConfigError("Failed to load config: " + str(err))
def from_dict(values): ''' Instantiate a BlockadeConfig instance based on a given dictionary of configuration values ''' try: containers = values['containers'] parsed_containers = {} for name, container_dict in containers.items(): try: # one config entry might result in many container # instances (indicated by the 'count' config value) for cnt in BlockadeContainerConfig.from_dict(name, container_dict): # check for duplicate 'container_name' definitions if cnt.container_name: cname = cnt.container_name existing = [c for c in parsed_containers.values() if c.container_name == cname] if existing: raise BlockadeConfigError("Duplicate 'container_name' definition: %s" % (cname)) parsed_containers[cnt.name] = cnt except Exception as err: raise BlockadeConfigError( "Container '%s' config problem: %s" % (name, err)) network = values.get('network') if network: defaults = _DEFAULT_NETWORK_CONFIG.copy() defaults.update(network) network = defaults else: network = _DEFAULT_NETWORK_CONFIG.copy() return BlockadeConfig(parsed_containers, network=network) except KeyError as err: raise BlockadeConfigError("Config missing value: " + str(err)) except Exception as err: # TODO log this to some debug stream? raise BlockadeConfigError("Failed to load config: " + str(err))
[ "Instantiate", "a", "BlockadeConfig", "instance", "based", "on", "a", "given", "dictionary", "of", "configuration", "values" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/config.py#L122-L161
[ "def", "from_dict", "(", "values", ")", ":", "try", ":", "containers", "=", "values", "[", "'containers'", "]", "parsed_containers", "=", "{", "}", "for", "name", ",", "container_dict", "in", "containers", ".", "items", "(", ")", ":", "try", ":", "# one config entry might result in many container", "# instances (indicated by the 'count' config value)", "for", "cnt", "in", "BlockadeContainerConfig", ".", "from_dict", "(", "name", ",", "container_dict", ")", ":", "# check for duplicate 'container_name' definitions", "if", "cnt", ".", "container_name", ":", "cname", "=", "cnt", ".", "container_name", "existing", "=", "[", "c", "for", "c", "in", "parsed_containers", ".", "values", "(", ")", "if", "c", ".", "container_name", "==", "cname", "]", "if", "existing", ":", "raise", "BlockadeConfigError", "(", "\"Duplicate 'container_name' definition: %s\"", "%", "(", "cname", ")", ")", "parsed_containers", "[", "cnt", ".", "name", "]", "=", "cnt", "except", "Exception", "as", "err", ":", "raise", "BlockadeConfigError", "(", "\"Container '%s' config problem: %s\"", "%", "(", "name", ",", "err", ")", ")", "network", "=", "values", ".", "get", "(", "'network'", ")", "if", "network", ":", "defaults", "=", "_DEFAULT_NETWORK_CONFIG", ".", "copy", "(", ")", "defaults", ".", "update", "(", "network", ")", "network", "=", "defaults", "else", ":", "network", "=", "_DEFAULT_NETWORK_CONFIG", ".", "copy", "(", ")", "return", "BlockadeConfig", "(", "parsed_containers", ",", "network", "=", "network", ")", "except", "KeyError", "as", "err", ":", "raise", "BlockadeConfigError", "(", "\"Config missing value: \"", "+", "str", "(", "err", ")", ")", "except", "Exception", "as", "err", ":", "# TODO log this to some debug stream?", "raise", "BlockadeConfigError", "(", "\"Failed to load config: \"", "+", "str", "(", "err", ")", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_up
Start the containers and link them together
blockade/cli.py
def cmd_up(opts): """Start the containers and link them together """ config = load_config(opts.config) b = get_blockade(config, opts) containers = b.create(verbose=opts.verbose, force=opts.force) print_containers(containers, opts.json)
def cmd_up(opts): """Start the containers and link them together """ config = load_config(opts.config) b = get_blockade(config, opts) containers = b.create(verbose=opts.verbose, force=opts.force) print_containers(containers, opts.json)
[ "Start", "the", "containers", "and", "link", "them", "together" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L160-L166
[ "def", "cmd_up", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "containers", "=", "b", ".", "create", "(", "verbose", "=", "opts", ".", "verbose", ",", "force", "=", "opts", ".", "force", ")", "print_containers", "(", "containers", ",", "opts", ".", "json", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_destroy
Destroy all containers and restore networks
blockade/cli.py
def cmd_destroy(opts): """Destroy all containers and restore networks """ config = load_config(opts.config) b = get_blockade(config, opts) b.destroy()
def cmd_destroy(opts): """Destroy all containers and restore networks """ config = load_config(opts.config) b = get_blockade(config, opts) b.destroy()
[ "Destroy", "all", "containers", "and", "restore", "networks" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L169-L174
[ "def", "cmd_destroy", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "b", ".", "destroy", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_status
Print status of containers and networks
blockade/cli.py
def cmd_status(opts): """Print status of containers and networks """ config = load_config(opts.config) b = get_blockade(config, opts) containers = b.status() print_containers(containers, opts.json)
def cmd_status(opts): """Print status of containers and networks """ config = load_config(opts.config) b = get_blockade(config, opts) containers = b.status() print_containers(containers, opts.json)
[ "Print", "status", "of", "containers", "and", "networks" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L177-L183
[ "def", "cmd_status", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "containers", "=", "b", ".", "status", "(", ")", "print_containers", "(", "containers", ",", "opts", ".", "json", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_kill
Kill some or all containers
blockade/cli.py
def cmd_kill(opts): """Kill some or all containers """ kill_signal = opts.signal if hasattr(opts, 'signal') else "SIGKILL" __with_containers(opts, Blockade.kill, signal=kill_signal)
def cmd_kill(opts): """Kill some or all containers """ kill_signal = opts.signal if hasattr(opts, 'signal') else "SIGKILL" __with_containers(opts, Blockade.kill, signal=kill_signal)
[ "Kill", "some", "or", "all", "containers" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L210-L214
[ "def", "cmd_kill", "(", "opts", ")", ":", "kill_signal", "=", "opts", ".", "signal", "if", "hasattr", "(", "opts", ",", "'signal'", ")", "else", "\"SIGKILL\"", "__with_containers", "(", "opts", ",", "Blockade", ".", "kill", ",", "signal", "=", "kill_signal", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_partition
Partition the network between containers Replaces any existing partitions outright. Any containers NOT specified in arguments will be globbed into a single implicit partition. For example if you have three containers: c1, c2, and c3 and you run: blockade partition c1 The result will be a partition with just c1 and another partition with c2 and c3. Alternatively, --random may be specified, and zero or more random partitions will be generated by blockade.
blockade/cli.py
def cmd_partition(opts): """Partition the network between containers Replaces any existing partitions outright. Any containers NOT specified in arguments will be globbed into a single implicit partition. For example if you have three containers: c1, c2, and c3 and you run: blockade partition c1 The result will be a partition with just c1 and another partition with c2 and c3. Alternatively, --random may be specified, and zero or more random partitions will be generated by blockade. """ config = load_config(opts.config) b = get_blockade(config, opts) if opts.random: if opts.partitions: raise BlockadeError("Either specify individual partitions " "or --random, but not both") b.random_partition() else: partitions = [] for partition in opts.partitions: names = [] for name in partition.split(","): name = name.strip() if name: names.append(name) partitions.append(names) if not partitions: raise BlockadeError("Either specify individual partitions " "or random") b.partition(partitions)
def cmd_partition(opts): """Partition the network between containers Replaces any existing partitions outright. Any containers NOT specified in arguments will be globbed into a single implicit partition. For example if you have three containers: c1, c2, and c3 and you run: blockade partition c1 The result will be a partition with just c1 and another partition with c2 and c3. Alternatively, --random may be specified, and zero or more random partitions will be generated by blockade. """ config = load_config(opts.config) b = get_blockade(config, opts) if opts.random: if opts.partitions: raise BlockadeError("Either specify individual partitions " "or --random, but not both") b.random_partition() else: partitions = [] for partition in opts.partitions: names = [] for name in partition.split(","): name = name.strip() if name: names.append(name) partitions.append(names) if not partitions: raise BlockadeError("Either specify individual partitions " "or random") b.partition(partitions)
[ "Partition", "the", "network", "between", "containers" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L304-L340
[ "def", "cmd_partition", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "if", "opts", ".", "random", ":", "if", "opts", ".", "partitions", ":", "raise", "BlockadeError", "(", "\"Either specify individual partitions \"", "\"or --random, but not both\"", ")", "b", ".", "random_partition", "(", ")", "else", ":", "partitions", "=", "[", "]", "for", "partition", "in", "opts", ".", "partitions", ":", "names", "=", "[", "]", "for", "name", "in", "partition", ".", "split", "(", "\",\"", ")", ":", "name", "=", "name", ".", "strip", "(", ")", "if", "name", ":", "names", ".", "append", "(", "name", ")", "partitions", ".", "append", "(", "names", ")", "if", "not", "partitions", ":", "raise", "BlockadeError", "(", "\"Either specify individual partitions \"", "\"or random\"", ")", "b", ".", "partition", "(", "partitions", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_join
Restore full networking between containers
blockade/cli.py
def cmd_join(opts): """Restore full networking between containers """ config = load_config(opts.config) b = get_blockade(config, opts) b.join()
def cmd_join(opts): """Restore full networking between containers """ config = load_config(opts.config) b = get_blockade(config, opts) b.join()
[ "Restore", "full", "networking", "between", "containers" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L343-L348
[ "def", "cmd_join", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "b", ".", "join", "(", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_logs
Fetch the logs of a container
blockade/cli.py
def cmd_logs(opts): """Fetch the logs of a container """ config = load_config(opts.config) b = get_blockade(config, opts) puts(b.logs(opts.container).decode(encoding='UTF-8'))
def cmd_logs(opts): """Fetch the logs of a container """ config = load_config(opts.config) b = get_blockade(config, opts) puts(b.logs(opts.container).decode(encoding='UTF-8'))
[ "Fetch", "the", "logs", "of", "a", "container" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L351-L356
[ "def", "cmd_logs", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "puts", "(", "b", ".", "logs", "(", "opts", ".", "container", ")", ".", "decode", "(", "encoding", "=", "'UTF-8'", ")", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_daemon
Start the Blockade REST API
blockade/cli.py
def cmd_daemon(opts): """Start the Blockade REST API """ if opts.data_dir is None: raise BlockadeError("You must supply a data directory for the daemon") rest.start(data_dir=opts.data_dir, port=opts.port, debug=opts.debug, host_exec=get_host_exec())
def cmd_daemon(opts): """Start the Blockade REST API """ if opts.data_dir is None: raise BlockadeError("You must supply a data directory for the daemon") rest.start(data_dir=opts.data_dir, port=opts.port, debug=opts.debug, host_exec=get_host_exec())
[ "Start", "the", "Blockade", "REST", "API" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L359-L365
[ "def", "cmd_daemon", "(", "opts", ")", ":", "if", "opts", ".", "data_dir", "is", "None", ":", "raise", "BlockadeError", "(", "\"You must supply a data directory for the daemon\"", ")", "rest", ".", "start", "(", "data_dir", "=", "opts", ".", "data_dir", ",", "port", "=", "opts", ".", "port", ",", "debug", "=", "opts", ".", "debug", ",", "host_exec", "=", "get_host_exec", "(", ")", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_add
Add one or more existing Docker containers to a Blockade group
blockade/cli.py
def cmd_add(opts): """Add one or more existing Docker containers to a Blockade group """ config = load_config(opts.config) b = get_blockade(config, opts) b.add_container(opts.containers)
def cmd_add(opts): """Add one or more existing Docker containers to a Blockade group """ config = load_config(opts.config) b = get_blockade(config, opts) b.add_container(opts.containers)
[ "Add", "one", "or", "more", "existing", "Docker", "containers", "to", "a", "Blockade", "group" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L368-L373
[ "def", "cmd_add", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "b", ".", "add_container", "(", "opts", ".", "containers", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
cmd_events
Get the event log for a given blockade
blockade/cli.py
def cmd_events(opts): """Get the event log for a given blockade """ config = load_config(opts.config) b = get_blockade(config, opts) if opts.json: outf = None _write = puts if opts.output is not None: outf = open(opts.output, "w") _write = outf.write try: delim = "" logs = b.get_audit().read_logs(as_json=False) _write('{"events": [') _write(os.linesep) for l in logs: _write(delim + l) delim = "," + os.linesep _write(os.linesep) _write(']}') finally: if opts.output is not None: outf.close() else: puts(colored.blue(columns(["EVENT", 10], ["TARGET", 16], ["STATUS", 8], ["TIME", 16], ["MESSAGE", 25]))) logs = b.get_audit().read_logs(as_json=True) for l in logs: puts(columns([l['event'], 10], [str([str(t) for t in l['targets']]), 16], [l['status'], 8], [str(l['timestamp']), 16], [l['message'], 25]))
def cmd_events(opts): """Get the event log for a given blockade """ config = load_config(opts.config) b = get_blockade(config, opts) if opts.json: outf = None _write = puts if opts.output is not None: outf = open(opts.output, "w") _write = outf.write try: delim = "" logs = b.get_audit().read_logs(as_json=False) _write('{"events": [') _write(os.linesep) for l in logs: _write(delim + l) delim = "," + os.linesep _write(os.linesep) _write(']}') finally: if opts.output is not None: outf.close() else: puts(colored.blue(columns(["EVENT", 10], ["TARGET", 16], ["STATUS", 8], ["TIME", 16], ["MESSAGE", 25]))) logs = b.get_audit().read_logs(as_json=True) for l in logs: puts(columns([l['event'], 10], [str([str(t) for t in l['targets']]), 16], [l['status'], 8], [str(l['timestamp']), 16], [l['message'], 25]))
[ "Get", "the", "event", "log", "for", "a", "given", "blockade" ]
worstcase/blockade
python
https://github.com/worstcase/blockade/blob/3dc6ad803f0b0d56586dec9542a6a06aa06cf569/blockade/cli.py#L383-L421
[ "def", "cmd_events", "(", "opts", ")", ":", "config", "=", "load_config", "(", "opts", ".", "config", ")", "b", "=", "get_blockade", "(", "config", ",", "opts", ")", "if", "opts", ".", "json", ":", "outf", "=", "None", "_write", "=", "puts", "if", "opts", ".", "output", "is", "not", "None", ":", "outf", "=", "open", "(", "opts", ".", "output", ",", "\"w\"", ")", "_write", "=", "outf", ".", "write", "try", ":", "delim", "=", "\"\"", "logs", "=", "b", ".", "get_audit", "(", ")", ".", "read_logs", "(", "as_json", "=", "False", ")", "_write", "(", "'{\"events\": ['", ")", "_write", "(", "os", ".", "linesep", ")", "for", "l", "in", "logs", ":", "_write", "(", "delim", "+", "l", ")", "delim", "=", "\",\"", "+", "os", ".", "linesep", "_write", "(", "os", ".", "linesep", ")", "_write", "(", "']}'", ")", "finally", ":", "if", "opts", ".", "output", "is", "not", "None", ":", "outf", ".", "close", "(", ")", "else", ":", "puts", "(", "colored", ".", "blue", "(", "columns", "(", "[", "\"EVENT\"", ",", "10", "]", ",", "[", "\"TARGET\"", ",", "16", "]", ",", "[", "\"STATUS\"", ",", "8", "]", ",", "[", "\"TIME\"", ",", "16", "]", ",", "[", "\"MESSAGE\"", ",", "25", "]", ")", ")", ")", "logs", "=", "b", ".", "get_audit", "(", ")", ".", "read_logs", "(", "as_json", "=", "True", ")", "for", "l", "in", "logs", ":", "puts", "(", "columns", "(", "[", "l", "[", "'event'", "]", ",", "10", "]", ",", "[", "str", "(", "[", "str", "(", "t", ")", "for", "t", "in", "l", "[", "'targets'", "]", "]", ")", ",", "16", "]", ",", "[", "l", "[", "'status'", "]", ",", "8", "]", ",", "[", "str", "(", "l", "[", "'timestamp'", "]", ")", ",", "16", "]", ",", "[", "l", "[", "'message'", "]", ",", "25", "]", ")", ")" ]
3dc6ad803f0b0d56586dec9542a6a06aa06cf569
valid
set_cors_headers
Performs the actual evaluation of Flas-CORS options and actually modifies the response object. This function is used both in the decorator and the after_request callback
flask_cors/core.py
def set_cors_headers(resp, options): """ Performs the actual evaluation of Flas-CORS options and actually modifies the response object. This function is used both in the decorator and the after_request callback """ # If CORS has already been evaluated via the decorator, skip if hasattr(resp, FLASK_CORS_EVALUATED): LOG.debug('CORS have been already evaluated, skipping') return resp # Some libraries, like OAuthlib, set resp.headers to non Multidict # objects (Werkzeug Headers work as well). This is a problem because # headers allow repeated values. if (not isinstance(resp.headers, Headers) and not isinstance(resp.headers, MultiDict)): resp.headers = MultiDict(resp.headers) headers_to_set = get_cors_headers(options, request.headers, request.method) LOG.debug('Settings CORS headers: %s', str(headers_to_set)) for k, v in headers_to_set.items(): resp.headers.add(k, v) return resp
def set_cors_headers(resp, options): """ Performs the actual evaluation of Flas-CORS options and actually modifies the response object. This function is used both in the decorator and the after_request callback """ # If CORS has already been evaluated via the decorator, skip if hasattr(resp, FLASK_CORS_EVALUATED): LOG.debug('CORS have been already evaluated, skipping') return resp # Some libraries, like OAuthlib, set resp.headers to non Multidict # objects (Werkzeug Headers work as well). This is a problem because # headers allow repeated values. if (not isinstance(resp.headers, Headers) and not isinstance(resp.headers, MultiDict)): resp.headers = MultiDict(resp.headers) headers_to_set = get_cors_headers(options, request.headers, request.method) LOG.debug('Settings CORS headers: %s', str(headers_to_set)) for k, v in headers_to_set.items(): resp.headers.add(k, v) return resp
[ "Performs", "the", "actual", "evaluation", "of", "Flas", "-", "CORS", "options", "and", "actually", "modifies", "the", "response", "object", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/core.py#L219-L247
[ "def", "set_cors_headers", "(", "resp", ",", "options", ")", ":", "# If CORS has already been evaluated via the decorator, skip", "if", "hasattr", "(", "resp", ",", "FLASK_CORS_EVALUATED", ")", ":", "LOG", ".", "debug", "(", "'CORS have been already evaluated, skipping'", ")", "return", "resp", "# Some libraries, like OAuthlib, set resp.headers to non Multidict", "# objects (Werkzeug Headers work as well). This is a problem because", "# headers allow repeated values.", "if", "(", "not", "isinstance", "(", "resp", ".", "headers", ",", "Headers", ")", "and", "not", "isinstance", "(", "resp", ".", "headers", ",", "MultiDict", ")", ")", ":", "resp", ".", "headers", "=", "MultiDict", "(", "resp", ".", "headers", ")", "headers_to_set", "=", "get_cors_headers", "(", "options", ",", "request", ".", "headers", ",", "request", ".", "method", ")", "LOG", ".", "debug", "(", "'Settings CORS headers: %s'", ",", "str", "(", "headers_to_set", ")", ")", "for", "k", ",", "v", "in", "headers_to_set", ".", "items", "(", ")", ":", "resp", ".", "headers", ".", "add", "(", "k", ",", "v", ")", "return", "resp" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
try_match
Safely attempts to match a pattern or string to a request origin.
flask_cors/core.py
def try_match(request_origin, maybe_regex): """Safely attempts to match a pattern or string to a request origin.""" if isinstance(maybe_regex, RegexObject): return re.match(maybe_regex, request_origin) elif probably_regex(maybe_regex): return re.match(maybe_regex, request_origin, flags=re.IGNORECASE) else: try: return request_origin.lower() == maybe_regex.lower() except AttributeError: return request_origin == maybe_regex
def try_match(request_origin, maybe_regex): """Safely attempts to match a pattern or string to a request origin.""" if isinstance(maybe_regex, RegexObject): return re.match(maybe_regex, request_origin) elif probably_regex(maybe_regex): return re.match(maybe_regex, request_origin, flags=re.IGNORECASE) else: try: return request_origin.lower() == maybe_regex.lower() except AttributeError: return request_origin == maybe_regex
[ "Safely", "attempts", "to", "match", "a", "pattern", "or", "string", "to", "a", "request", "origin", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/core.py#L270-L280
[ "def", "try_match", "(", "request_origin", ",", "maybe_regex", ")", ":", "if", "isinstance", "(", "maybe_regex", ",", "RegexObject", ")", ":", "return", "re", ".", "match", "(", "maybe_regex", ",", "request_origin", ")", "elif", "probably_regex", "(", "maybe_regex", ")", ":", "return", "re", ".", "match", "(", "maybe_regex", ",", "request_origin", ",", "flags", "=", "re", ".", "IGNORECASE", ")", "else", ":", "try", ":", "return", "request_origin", ".", "lower", "(", ")", "==", "maybe_regex", ".", "lower", "(", ")", "except", "AttributeError", ":", "return", "request_origin", "==", "maybe_regex" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
get_cors_options
Compute CORS options for an application by combining the DEFAULT_OPTIONS, the app's configuration-specified options and any dictionaries passed. The last specified option wins.
flask_cors/core.py
def get_cors_options(appInstance, *dicts): """ Compute CORS options for an application by combining the DEFAULT_OPTIONS, the app's configuration-specified options and any dictionaries passed. The last specified option wins. """ options = DEFAULT_OPTIONS.copy() options.update(get_app_kwarg_dict(appInstance)) if dicts: for d in dicts: options.update(d) return serialize_options(options)
def get_cors_options(appInstance, *dicts): """ Compute CORS options for an application by combining the DEFAULT_OPTIONS, the app's configuration-specified options and any dictionaries passed. The last specified option wins. """ options = DEFAULT_OPTIONS.copy() options.update(get_app_kwarg_dict(appInstance)) if dicts: for d in dicts: options.update(d) return serialize_options(options)
[ "Compute", "CORS", "options", "for", "an", "application", "by", "combining", "the", "DEFAULT_OPTIONS", "the", "app", "s", "configuration", "-", "specified", "options", "and", "any", "dictionaries", "passed", ".", "The", "last", "specified", "option", "wins", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/core.py#L283-L295
[ "def", "get_cors_options", "(", "appInstance", ",", "*", "dicts", ")", ":", "options", "=", "DEFAULT_OPTIONS", ".", "copy", "(", ")", "options", ".", "update", "(", "get_app_kwarg_dict", "(", "appInstance", ")", ")", "if", "dicts", ":", "for", "d", "in", "dicts", ":", "options", ".", "update", "(", "d", ")", "return", "serialize_options", "(", "options", ")" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
get_app_kwarg_dict
Returns the dictionary of CORS specific app configurations.
flask_cors/core.py
def get_app_kwarg_dict(appInstance=None): """Returns the dictionary of CORS specific app configurations.""" app = (appInstance or current_app) # In order to support blueprints which do not have a config attribute app_config = getattr(app, 'config', {}) return { k.lower().replace('cors_', ''): app_config.get(k) for k in CONFIG_OPTIONS if app_config.get(k) is not None }
def get_app_kwarg_dict(appInstance=None): """Returns the dictionary of CORS specific app configurations.""" app = (appInstance or current_app) # In order to support blueprints which do not have a config attribute app_config = getattr(app, 'config', {}) return { k.lower().replace('cors_', ''): app_config.get(k) for k in CONFIG_OPTIONS if app_config.get(k) is not None }
[ "Returns", "the", "dictionary", "of", "CORS", "specific", "app", "configurations", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/core.py#L298-L309
[ "def", "get_app_kwarg_dict", "(", "appInstance", "=", "None", ")", ":", "app", "=", "(", "appInstance", "or", "current_app", ")", "# In order to support blueprints which do not have a config attribute", "app_config", "=", "getattr", "(", "app", ",", "'config'", ",", "{", "}", ")", "return", "{", "k", ".", "lower", "(", ")", ".", "replace", "(", "'cors_'", ",", "''", ")", ":", "app_config", ".", "get", "(", "k", ")", "for", "k", "in", "CONFIG_OPTIONS", "if", "app_config", ".", "get", "(", "k", ")", "is", "not", "None", "}" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
ensure_iterable
Wraps scalars or string types as a list, or returns the iterable instance.
flask_cors/core.py
def ensure_iterable(inst): """ Wraps scalars or string types as a list, or returns the iterable instance. """ if isinstance(inst, string_types): return [inst] elif not isinstance(inst, collections.Iterable): return [inst] else: return inst
def ensure_iterable(inst): """ Wraps scalars or string types as a list, or returns the iterable instance. """ if isinstance(inst, string_types): return [inst] elif not isinstance(inst, collections.Iterable): return [inst] else: return inst
[ "Wraps", "scalars", "or", "string", "types", "as", "a", "list", "or", "returns", "the", "iterable", "instance", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/core.py#L334-L343
[ "def", "ensure_iterable", "(", "inst", ")", ":", "if", "isinstance", "(", "inst", ",", "string_types", ")", ":", "return", "[", "inst", "]", "elif", "not", "isinstance", "(", "inst", ",", "collections", ".", "Iterable", ")", ":", "return", "[", "inst", "]", "else", ":", "return", "inst" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
serialize_options
A helper method to serialize and processes the options dictionary.
flask_cors/core.py
def serialize_options(opts): """ A helper method to serialize and processes the options dictionary. """ options = (opts or {}).copy() for key in opts.keys(): if key not in DEFAULT_OPTIONS: LOG.warning("Unknown option passed to Flask-CORS: %s", key) # Ensure origins is a list of allowed origins with at least one entry. options['origins'] = sanitize_regex_param(options.get('origins')) options['allow_headers'] = sanitize_regex_param(options.get('allow_headers')) # This is expressly forbidden by the spec. Raise a value error so people # don't get burned in production. if r'.*' in options['origins'] and options['supports_credentials'] and options['send_wildcard']: raise ValueError("Cannot use supports_credentials in conjunction with" "an origin string of '*'. See: " "http://www.w3.org/TR/cors/#resource-requests") serialize_option(options, 'expose_headers') serialize_option(options, 'methods', upper=True) if isinstance(options.get('max_age'), timedelta): options['max_age'] = str(int(options['max_age'].total_seconds())) return options
def serialize_options(opts): """ A helper method to serialize and processes the options dictionary. """ options = (opts or {}).copy() for key in opts.keys(): if key not in DEFAULT_OPTIONS: LOG.warning("Unknown option passed to Flask-CORS: %s", key) # Ensure origins is a list of allowed origins with at least one entry. options['origins'] = sanitize_regex_param(options.get('origins')) options['allow_headers'] = sanitize_regex_param(options.get('allow_headers')) # This is expressly forbidden by the spec. Raise a value error so people # don't get burned in production. if r'.*' in options['origins'] and options['supports_credentials'] and options['send_wildcard']: raise ValueError("Cannot use supports_credentials in conjunction with" "an origin string of '*'. See: " "http://www.w3.org/TR/cors/#resource-requests") serialize_option(options, 'expose_headers') serialize_option(options, 'methods', upper=True) if isinstance(options.get('max_age'), timedelta): options['max_age'] = str(int(options['max_age'].total_seconds())) return options
[ "A", "helper", "method", "to", "serialize", "and", "processes", "the", "options", "dictionary", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/core.py#L349-L378
[ "def", "serialize_options", "(", "opts", ")", ":", "options", "=", "(", "opts", "or", "{", "}", ")", ".", "copy", "(", ")", "for", "key", "in", "opts", ".", "keys", "(", ")", ":", "if", "key", "not", "in", "DEFAULT_OPTIONS", ":", "LOG", ".", "warning", "(", "\"Unknown option passed to Flask-CORS: %s\"", ",", "key", ")", "# Ensure origins is a list of allowed origins with at least one entry.", "options", "[", "'origins'", "]", "=", "sanitize_regex_param", "(", "options", ".", "get", "(", "'origins'", ")", ")", "options", "[", "'allow_headers'", "]", "=", "sanitize_regex_param", "(", "options", ".", "get", "(", "'allow_headers'", ")", ")", "# This is expressly forbidden by the spec. Raise a value error so people", "# don't get burned in production.", "if", "r'.*'", "in", "options", "[", "'origins'", "]", "and", "options", "[", "'supports_credentials'", "]", "and", "options", "[", "'send_wildcard'", "]", ":", "raise", "ValueError", "(", "\"Cannot use supports_credentials in conjunction with\"", "\"an origin string of '*'. See: \"", "\"http://www.w3.org/TR/cors/#resource-requests\"", ")", "serialize_option", "(", "options", ",", "'expose_headers'", ")", "serialize_option", "(", "options", ",", "'methods'", ",", "upper", "=", "True", ")", "if", "isinstance", "(", "options", ".", "get", "(", "'max_age'", ")", ",", "timedelta", ")", ":", "options", "[", "'max_age'", "]", "=", "str", "(", "int", "(", "options", "[", "'max_age'", "]", ".", "total_seconds", "(", ")", ")", ")", "return", "options" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
cross_origin
This function is the decorator which is used to wrap a Flask route with. In the simplest case, simply use the default parameters to allow all origins in what is the most permissive configuration. If this method modifies state or performs authentication which may be brute-forced, you should add some degree of protection, such as Cross Site Forgery Request protection. :param origins: The origin, or list of origins to allow requests from. The origin(s) may be regular expressions, case-sensitive strings, or else an asterisk Default : '*' :type origins: list, string or regex :param methods: The method or list of methods which the allowed origins are allowed to access for non-simple requests. Default : [GET, HEAD, POST, OPTIONS, PUT, PATCH, DELETE] :type methods: list or string :param expose_headers: The header or list which are safe to expose to the API of a CORS API specification. Default : None :type expose_headers: list or string :param allow_headers: The header or list of header field names which can be used when this resource is accessed by allowed origins. The header(s) may be regular expressions, case-sensitive strings, or else an asterisk. Default : '*', allow all headers :type allow_headers: list, string or regex :param supports_credentials: Allows users to make authenticated requests. If true, injects the `Access-Control-Allow-Credentials` header in responses. This allows cookies and credentials to be submitted across domains. :note: This option cannot be used in conjuction with a '*' origin Default : False :type supports_credentials: bool :param max_age: The maximum time for which this CORS request maybe cached. This value is set as the `Access-Control-Max-Age` header. Default : None :type max_age: timedelta, integer, string or None :param send_wildcard: If True, and the origins parameter is `*`, a wildcard `Access-Control-Allow-Origin` header is sent, rather than the request's `Origin` header. Default : False :type send_wildcard: bool :param vary_header: If True, the header Vary: Origin will be returned as per the W3 implementation guidelines. Setting this header when the `Access-Control-Allow-Origin` is dynamically generated (e.g. when there is more than one allowed origin, and an Origin than '*' is returned) informs CDNs and other caches that the CORS headers are dynamic, and cannot be cached. If False, the Vary header will never be injected or altered. Default : True :type vary_header: bool :param automatic_options: Only applies to the `cross_origin` decorator. If True, Flask-CORS will override Flask's default OPTIONS handling to return CORS headers for OPTIONS requests. Default : True :type automatic_options: bool
flask_cors/decorator.py
def cross_origin(*args, **kwargs): """ This function is the decorator which is used to wrap a Flask route with. In the simplest case, simply use the default parameters to allow all origins in what is the most permissive configuration. If this method modifies state or performs authentication which may be brute-forced, you should add some degree of protection, such as Cross Site Forgery Request protection. :param origins: The origin, or list of origins to allow requests from. The origin(s) may be regular expressions, case-sensitive strings, or else an asterisk Default : '*' :type origins: list, string or regex :param methods: The method or list of methods which the allowed origins are allowed to access for non-simple requests. Default : [GET, HEAD, POST, OPTIONS, PUT, PATCH, DELETE] :type methods: list or string :param expose_headers: The header or list which are safe to expose to the API of a CORS API specification. Default : None :type expose_headers: list or string :param allow_headers: The header or list of header field names which can be used when this resource is accessed by allowed origins. The header(s) may be regular expressions, case-sensitive strings, or else an asterisk. Default : '*', allow all headers :type allow_headers: list, string or regex :param supports_credentials: Allows users to make authenticated requests. If true, injects the `Access-Control-Allow-Credentials` header in responses. This allows cookies and credentials to be submitted across domains. :note: This option cannot be used in conjuction with a '*' origin Default : False :type supports_credentials: bool :param max_age: The maximum time for which this CORS request maybe cached. This value is set as the `Access-Control-Max-Age` header. Default : None :type max_age: timedelta, integer, string or None :param send_wildcard: If True, and the origins parameter is `*`, a wildcard `Access-Control-Allow-Origin` header is sent, rather than the request's `Origin` header. Default : False :type send_wildcard: bool :param vary_header: If True, the header Vary: Origin will be returned as per the W3 implementation guidelines. Setting this header when the `Access-Control-Allow-Origin` is dynamically generated (e.g. when there is more than one allowed origin, and an Origin than '*' is returned) informs CDNs and other caches that the CORS headers are dynamic, and cannot be cached. If False, the Vary header will never be injected or altered. Default : True :type vary_header: bool :param automatic_options: Only applies to the `cross_origin` decorator. If True, Flask-CORS will override Flask's default OPTIONS handling to return CORS headers for OPTIONS requests. Default : True :type automatic_options: bool """ _options = kwargs def decorator(f): LOG.debug("Enabling %s for cross_origin using options:%s", f, _options) # If True, intercept OPTIONS requests by modifying the view function, # replicating Flask's default behavior, and wrapping the response with # CORS headers. # # If f.provide_automatic_options is unset or True, Flask's route # decorator (which is actually wraps the function object we return) # intercepts OPTIONS handling, and requests will not have CORS headers if _options.get('automatic_options', True): f.required_methods = getattr(f, 'required_methods', set()) f.required_methods.add('OPTIONS') f.provide_automatic_options = False def wrapped_function(*args, **kwargs): # Handle setting of Flask-Cors parameters options = get_cors_options(current_app, _options) if options.get('automatic_options') and request.method == 'OPTIONS': resp = current_app.make_default_options_response() else: resp = make_response(f(*args, **kwargs)) set_cors_headers(resp, options) setattr(resp, FLASK_CORS_EVALUATED, True) return resp return update_wrapper(wrapped_function, f) return decorator
def cross_origin(*args, **kwargs): """ This function is the decorator which is used to wrap a Flask route with. In the simplest case, simply use the default parameters to allow all origins in what is the most permissive configuration. If this method modifies state or performs authentication which may be brute-forced, you should add some degree of protection, such as Cross Site Forgery Request protection. :param origins: The origin, or list of origins to allow requests from. The origin(s) may be regular expressions, case-sensitive strings, or else an asterisk Default : '*' :type origins: list, string or regex :param methods: The method or list of methods which the allowed origins are allowed to access for non-simple requests. Default : [GET, HEAD, POST, OPTIONS, PUT, PATCH, DELETE] :type methods: list or string :param expose_headers: The header or list which are safe to expose to the API of a CORS API specification. Default : None :type expose_headers: list or string :param allow_headers: The header or list of header field names which can be used when this resource is accessed by allowed origins. The header(s) may be regular expressions, case-sensitive strings, or else an asterisk. Default : '*', allow all headers :type allow_headers: list, string or regex :param supports_credentials: Allows users to make authenticated requests. If true, injects the `Access-Control-Allow-Credentials` header in responses. This allows cookies and credentials to be submitted across domains. :note: This option cannot be used in conjuction with a '*' origin Default : False :type supports_credentials: bool :param max_age: The maximum time for which this CORS request maybe cached. This value is set as the `Access-Control-Max-Age` header. Default : None :type max_age: timedelta, integer, string or None :param send_wildcard: If True, and the origins parameter is `*`, a wildcard `Access-Control-Allow-Origin` header is sent, rather than the request's `Origin` header. Default : False :type send_wildcard: bool :param vary_header: If True, the header Vary: Origin will be returned as per the W3 implementation guidelines. Setting this header when the `Access-Control-Allow-Origin` is dynamically generated (e.g. when there is more than one allowed origin, and an Origin than '*' is returned) informs CDNs and other caches that the CORS headers are dynamic, and cannot be cached. If False, the Vary header will never be injected or altered. Default : True :type vary_header: bool :param automatic_options: Only applies to the `cross_origin` decorator. If True, Flask-CORS will override Flask's default OPTIONS handling to return CORS headers for OPTIONS requests. Default : True :type automatic_options: bool """ _options = kwargs def decorator(f): LOG.debug("Enabling %s for cross_origin using options:%s", f, _options) # If True, intercept OPTIONS requests by modifying the view function, # replicating Flask's default behavior, and wrapping the response with # CORS headers. # # If f.provide_automatic_options is unset or True, Flask's route # decorator (which is actually wraps the function object we return) # intercepts OPTIONS handling, and requests will not have CORS headers if _options.get('automatic_options', True): f.required_methods = getattr(f, 'required_methods', set()) f.required_methods.add('OPTIONS') f.provide_automatic_options = False def wrapped_function(*args, **kwargs): # Handle setting of Flask-Cors parameters options = get_cors_options(current_app, _options) if options.get('automatic_options') and request.method == 'OPTIONS': resp = current_app.make_default_options_response() else: resp = make_response(f(*args, **kwargs)) set_cors_headers(resp, options) setattr(resp, FLASK_CORS_EVALUATED, True) return resp return update_wrapper(wrapped_function, f) return decorator
[ "This", "function", "is", "the", "decorator", "which", "is", "used", "to", "wrap", "a", "Flask", "route", "with", ".", "In", "the", "simplest", "case", "simply", "use", "the", "default", "parameters", "to", "allow", "all", "origins", "in", "what", "is", "the", "most", "permissive", "configuration", ".", "If", "this", "method", "modifies", "state", "or", "performs", "authentication", "which", "may", "be", "brute", "-", "forced", "you", "should", "add", "some", "degree", "of", "protection", "such", "as", "Cross", "Site", "Forgery", "Request", "protection", "." ]
corydolphin/flask-cors
python
https://github.com/corydolphin/flask-cors/blob/13fbb1ea4c1bb422de91a726c3c7f1038d3743a3/flask_cors/decorator.py#L18-L135
[ "def", "cross_origin", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "_options", "=", "kwargs", "def", "decorator", "(", "f", ")", ":", "LOG", ".", "debug", "(", "\"Enabling %s for cross_origin using options:%s\"", ",", "f", ",", "_options", ")", "# If True, intercept OPTIONS requests by modifying the view function,", "# replicating Flask's default behavior, and wrapping the response with", "# CORS headers.", "#", "# If f.provide_automatic_options is unset or True, Flask's route", "# decorator (which is actually wraps the function object we return)", "# intercepts OPTIONS handling, and requests will not have CORS headers", "if", "_options", ".", "get", "(", "'automatic_options'", ",", "True", ")", ":", "f", ".", "required_methods", "=", "getattr", "(", "f", ",", "'required_methods'", ",", "set", "(", ")", ")", "f", ".", "required_methods", ".", "add", "(", "'OPTIONS'", ")", "f", ".", "provide_automatic_options", "=", "False", "def", "wrapped_function", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "# Handle setting of Flask-Cors parameters", "options", "=", "get_cors_options", "(", "current_app", ",", "_options", ")", "if", "options", ".", "get", "(", "'automatic_options'", ")", "and", "request", ".", "method", "==", "'OPTIONS'", ":", "resp", "=", "current_app", ".", "make_default_options_response", "(", ")", "else", ":", "resp", "=", "make_response", "(", "f", "(", "*", "args", ",", "*", "*", "kwargs", ")", ")", "set_cors_headers", "(", "resp", ",", "options", ")", "setattr", "(", "resp", ",", "FLASK_CORS_EVALUATED", ",", "True", ")", "return", "resp", "return", "update_wrapper", "(", "wrapped_function", ",", "f", ")", "return", "decorator" ]
13fbb1ea4c1bb422de91a726c3c7f1038d3743a3
valid
calendar
This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: type (string); "holiday" or "trade" direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result
pyEX/refdata.py
def calendar(type='holiday', direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: type (string); "holiday" or "trade" direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' if startDate: startDate = _strOrDate(startDate) return _getJson('ref-data/us/dates/{type}/{direction}/{last}/{date}'.format(type=type, direction=direction, last=last, date=startDate), token, version) return _getJson('ref-data/us/dates/' + type + '/' + direction + '/' + str(last), token, version)
def calendar(type='holiday', direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: type (string); "holiday" or "trade" direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' if startDate: startDate = _strOrDate(startDate) return _getJson('ref-data/us/dates/{type}/{direction}/{last}/{date}'.format(type=type, direction=direction, last=last, date=startDate), token, version) return _getJson('ref-data/us/dates/' + type + '/' + direction + '/' + str(last), token, version)
[ "This", "call", "allows", "you", "to", "fetch", "a", "number", "of", "trade", "dates", "or", "holidays", "from", "a", "given", "date", ".", "For", "example", "if", "you", "want", "the", "next", "trading", "day", "you", "would", "call", "/", "ref", "-", "data", "/", "us", "/", "dates", "/", "trade", "/", "next", "/", "1", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L38-L58
[ "def", "calendar", "(", "type", "=", "'holiday'", ",", "direction", "=", "'next'", ",", "last", "=", "1", ",", "startDate", "=", "None", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "if", "startDate", ":", "startDate", "=", "_strOrDate", "(", "startDate", ")", "return", "_getJson", "(", "'ref-data/us/dates/{type}/{direction}/{last}/{date}'", ".", "format", "(", "type", "=", "type", ",", "direction", "=", "direction", ",", "last", "=", "last", ",", "date", "=", "startDate", ")", ",", "token", ",", "version", ")", "return", "_getJson", "(", "'ref-data/us/dates/'", "+", "type", "+", "'/'", "+", "direction", "+", "'/'", "+", "str", "(", "last", ")", ",", "token", ",", "version", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
calendarDF
This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: type (string); "holiday" or "trade" direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result
pyEX/refdata.py
def calendarDF(type='holiday', direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: type (string); "holiday" or "trade" direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' dat = pd.DataFrame(calendar(type, direction, last, startDate, token, version)) _toDatetime(dat) return dat
def calendarDF(type='holiday', direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: type (string); "holiday" or "trade" direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' dat = pd.DataFrame(calendar(type, direction, last, startDate, token, version)) _toDatetime(dat) return dat
[ "This", "call", "allows", "you", "to", "fetch", "a", "number", "of", "trade", "dates", "or", "holidays", "from", "a", "given", "date", ".", "For", "example", "if", "you", "want", "the", "next", "trading", "day", "you", "would", "call", "/", "ref", "-", "data", "/", "us", "/", "dates", "/", "trade", "/", "next", "/", "1", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L61-L80
[ "def", "calendarDF", "(", "type", "=", "'holiday'", ",", "direction", "=", "'next'", ",", "last", "=", "1", ",", "startDate", "=", "None", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "dat", "=", "pd", ".", "DataFrame", "(", "calendar", "(", "type", ",", "direction", ",", "last", ",", "startDate", ",", "token", ",", "version", ")", ")", "_toDatetime", "(", "dat", ")", "return", "dat" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
holidays
This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result
pyEX/refdata.py
def holidays(direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' return calendar('holiday', direction, last, startDate, token, version)
def holidays(direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' return calendar('holiday', direction, last, startDate, token, version)
[ "This", "call", "allows", "you", "to", "fetch", "a", "number", "of", "trade", "dates", "or", "holidays", "from", "a", "given", "date", ".", "For", "example", "if", "you", "want", "the", "next", "trading", "day", "you", "would", "call", "/", "ref", "-", "data", "/", "us", "/", "dates", "/", "trade", "/", "next", "/", "1", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L83-L99
[ "def", "holidays", "(", "direction", "=", "'next'", ",", "last", "=", "1", ",", "startDate", "=", "None", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "return", "calendar", "(", "'holiday'", ",", "direction", ",", "last", ",", "startDate", ",", "token", ",", "version", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
holidaysDF
This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result
pyEX/refdata.py
def holidaysDF(direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' return calendarDF('holiday', direction, last, startDate, token, version)
def holidaysDF(direction='next', last=1, startDate=None, token='', version=''): '''This call allows you to fetch a number of trade dates or holidays from a given date. For example, if you want the next trading day, you would call /ref-data/us/dates/trade/next/1. https://iexcloud.io/docs/api/#u-s-exchanges 8am, 9am, 12pm, 1pm UTC daily Args: direction (string); "next" or "last" last (int); number to move in direction startDate (date); start date for next or last, YYYYMMDD token (string); Access token version (string); API version Returns: dict: result ''' return calendarDF('holiday', direction, last, startDate, token, version)
[ "This", "call", "allows", "you", "to", "fetch", "a", "number", "of", "trade", "dates", "or", "holidays", "from", "a", "given", "date", ".", "For", "example", "if", "you", "want", "the", "next", "trading", "day", "you", "would", "call", "/", "ref", "-", "data", "/", "us", "/", "dates", "/", "trade", "/", "next", "/", "1", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L102-L118
[ "def", "holidaysDF", "(", "direction", "=", "'next'", ",", "last", "=", "1", ",", "startDate", "=", "None", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "return", "calendarDF", "(", "'holiday'", ",", "direction", ",", "last", ",", "startDate", ",", "token", ",", "version", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
internationalSymbols
This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: dict: result
pyEX/refdata.py
def internationalSymbols(region='', exchange='', token='', version=''): '''This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: dict: result ''' if region: return _getJson('ref-data/region/{region}/symbols'.format(region=region), token, version) elif exchange: return _getJson('ref-data/exchange/{exchange}/exchange'.format(exchange=exchange), token, version) return _getJson('ref-data/region/us/symbols', token, version)
def internationalSymbols(region='', exchange='', token='', version=''): '''This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: dict: result ''' if region: return _getJson('ref-data/region/{region}/symbols'.format(region=region), token, version) elif exchange: return _getJson('ref-data/exchange/{exchange}/exchange'.format(exchange=exchange), token, version) return _getJson('ref-data/region/us/symbols', token, version)
[ "This", "call", "returns", "an", "array", "of", "international", "symbols", "that", "IEX", "Cloud", "supports", "for", "API", "calls", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L186-L205
[ "def", "internationalSymbols", "(", "region", "=", "''", ",", "exchange", "=", "''", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "if", "region", ":", "return", "_getJson", "(", "'ref-data/region/{region}/symbols'", ".", "format", "(", "region", "=", "region", ")", ",", "token", ",", "version", ")", "elif", "exchange", ":", "return", "_getJson", "(", "'ref-data/exchange/{exchange}/exchange'", ".", "format", "(", "exchange", "=", "exchange", ")", ",", "token", ",", "version", ")", "return", "_getJson", "(", "'ref-data/region/us/symbols'", ",", "token", ",", "version", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
symbolsDF
This call returns an array of symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: dataframe: result
pyEX/refdata.py
def symbolsDF(token='', version=''): '''This call returns an array of symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: dataframe: result ''' df = pd.DataFrame(symbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
def symbolsDF(token='', version=''): '''This call returns an array of symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: dataframe: result ''' df = pd.DataFrame(symbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
[ "This", "call", "returns", "an", "array", "of", "symbols", "that", "IEX", "Cloud", "supports", "for", "API", "calls", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L208-L224
[ "def", "symbolsDF", "(", "token", "=", "''", ",", "version", "=", "''", ")", ":", "df", "=", "pd", ".", "DataFrame", "(", "symbols", "(", "token", ",", "version", ")", ")", "_toDatetime", "(", "df", ")", "_reindex", "(", "df", ",", "'symbol'", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
iexSymbolsDF
This call returns an array of symbols the Investors Exchange supports for trading. This list is updated daily as of 7:45 a.m. ET. Symbols may be added or removed by the Investors Exchange after the list was produced. https://iexcloud.io/docs/api/#iex-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result
pyEX/refdata.py
def iexSymbolsDF(token='', version=''): '''This call returns an array of symbols the Investors Exchange supports for trading. This list is updated daily as of 7:45 a.m. ET. Symbols may be added or removed by the Investors Exchange after the list was produced. https://iexcloud.io/docs/api/#iex-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(iexSymbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
def iexSymbolsDF(token='', version=''): '''This call returns an array of symbols the Investors Exchange supports for trading. This list is updated daily as of 7:45 a.m. ET. Symbols may be added or removed by the Investors Exchange after the list was produced. https://iexcloud.io/docs/api/#iex-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(iexSymbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
[ "This", "call", "returns", "an", "array", "of", "symbols", "the", "Investors", "Exchange", "supports", "for", "trading", ".", "This", "list", "is", "updated", "daily", "as", "of", "7", ":", "45", "a", ".", "m", ".", "ET", ".", "Symbols", "may", "be", "added", "or", "removed", "by", "the", "Investors", "Exchange", "after", "the", "list", "was", "produced", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L227-L244
[ "def", "iexSymbolsDF", "(", "token", "=", "''", ",", "version", "=", "''", ")", ":", "df", "=", "pd", ".", "DataFrame", "(", "iexSymbols", "(", "token", ",", "version", ")", ")", "_toDatetime", "(", "df", ")", "_reindex", "(", "df", ",", "'symbol'", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
mutualFundSymbolsDF
This call returns an array of mutual fund symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#mutual-fund-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result
pyEX/refdata.py
def mutualFundSymbolsDF(token='', version=''): '''This call returns an array of mutual fund symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#mutual-fund-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(mutualFundSymbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
def mutualFundSymbolsDF(token='', version=''): '''This call returns an array of mutual fund symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#mutual-fund-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(mutualFundSymbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
[ "This", "call", "returns", "an", "array", "of", "mutual", "fund", "symbols", "that", "IEX", "Cloud", "supports", "for", "API", "calls", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L247-L263
[ "def", "mutualFundSymbolsDF", "(", "token", "=", "''", ",", "version", "=", "''", ")", ":", "df", "=", "pd", ".", "DataFrame", "(", "mutualFundSymbols", "(", "token", ",", "version", ")", ")", "_toDatetime", "(", "df", ")", "_reindex", "(", "df", ",", "'symbol'", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
otcSymbolsDF
This call returns an array of OTC symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#otc-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result
pyEX/refdata.py
def otcSymbolsDF(token='', version=''): '''This call returns an array of OTC symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#otc-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(otcSymbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
def otcSymbolsDF(token='', version=''): '''This call returns an array of OTC symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#otc-symbols 8am, 9am, 12pm, 1pm UTC daily Args: token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(otcSymbols(token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
[ "This", "call", "returns", "an", "array", "of", "OTC", "symbols", "that", "IEX", "Cloud", "supports", "for", "API", "calls", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L266-L282
[ "def", "otcSymbolsDF", "(", "token", "=", "''", ",", "version", "=", "''", ")", ":", "df", "=", "pd", ".", "DataFrame", "(", "otcSymbols", "(", "token", ",", "version", ")", ")", "_toDatetime", "(", "df", ")", "_reindex", "(", "df", ",", "'symbol'", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
internationalSymbolsDF
This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: DataFrame: result
pyEX/refdata.py
def internationalSymbolsDF(region='', exchange='', token='', version=''): '''This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(internationalSymbols(region, exchange, token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
def internationalSymbolsDF(region='', exchange='', token='', version=''): '''This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: DataFrame: result ''' df = pd.DataFrame(internationalSymbols(region, exchange, token, version)) _toDatetime(df) _reindex(df, 'symbol') return df
[ "This", "call", "returns", "an", "array", "of", "international", "symbols", "that", "IEX", "Cloud", "supports", "for", "API", "calls", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L285-L303
[ "def", "internationalSymbolsDF", "(", "region", "=", "''", ",", "exchange", "=", "''", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "df", "=", "pd", ".", "DataFrame", "(", "internationalSymbols", "(", "region", ",", "exchange", ",", "token", ",", "version", ")", ")", "_toDatetime", "(", "df", ")", "_reindex", "(", "df", ",", "'symbol'", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
internationalSymbolsList
This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: list: result
pyEX/refdata.py
def internationalSymbolsList(region='', exchange='', token='', version=''): '''This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: list: result ''' return internationalSymbolsDF(region, exchange, token, version).index.tolist()
def internationalSymbolsList(region='', exchange='', token='', version=''): '''This call returns an array of international symbols that IEX Cloud supports for API calls. https://iexcloud.io/docs/api/#international-symbols 8am, 9am, 12pm, 1pm UTC daily Args: region (string); region, 2 letter case insensitive string of country codes using ISO 3166-1 alpha-2 exchange (string): Case insensitive string of Exchange using IEX Supported Exchanges list token (string); Access token version (string); API version Returns: list: result ''' return internationalSymbolsDF(region, exchange, token, version).index.tolist()
[ "This", "call", "returns", "an", "array", "of", "international", "symbols", "that", "IEX", "Cloud", "supports", "for", "API", "calls", "." ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/refdata.py#L371-L386
[ "def", "internationalSymbolsList", "(", "region", "=", "''", ",", "exchange", "=", "''", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "return", "internationalSymbolsDF", "(", "region", ",", "exchange", ",", "token", ",", "version", ")", ".", "index", ".", "tolist", "(", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
marketsDF
https://iextrading.com/developer/docs/#intraday
pyEX/markets.py
def marketsDF(token='', version=''): '''https://iextrading.com/developer/docs/#intraday''' df = pd.DataFrame(markets(token, version)) _toDatetime(df) return df
def marketsDF(token='', version=''): '''https://iextrading.com/developer/docs/#intraday''' df = pd.DataFrame(markets(token, version)) _toDatetime(df) return df
[ "https", ":", "//", "iextrading", ".", "com", "/", "developer", "/", "docs", "/", "#intraday" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/markets.py#L13-L17
[ "def", "marketsDF", "(", "token", "=", "''", ",", "version", "=", "''", ")", ":", "df", "=", "pd", ".", "DataFrame", "(", "markets", "(", "token", ",", "version", ")", ")", "_toDatetime", "(", "df", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_getJson
for backwards compat, accepting token and version but ignoring
pyEX/common.py
def _getJson(url, token='', version=''): '''for backwards compat, accepting token and version but ignoring''' if token: return _getJsonIEXCloud(url, token, version) return _getJsonOrig(url)
def _getJson(url, token='', version=''): '''for backwards compat, accepting token and version but ignoring''' if token: return _getJsonIEXCloud(url, token, version) return _getJsonOrig(url)
[ "for", "backwards", "compat", "accepting", "token", "and", "version", "but", "ignoring" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L91-L95
[ "def", "_getJson", "(", "url", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "if", "token", ":", "return", "_getJsonIEXCloud", "(", "url", ",", "token", ",", "version", ")", "return", "_getJsonOrig", "(", "url", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_getJsonOrig
internal
pyEX/common.py
def _getJsonOrig(url): '''internal''' url = _URL_PREFIX + url resp = requests.get(urlparse(url).geturl(), proxies=_PYEX_PROXIES) if resp.status_code == 200: return resp.json() raise PyEXception('Response %d - ' % resp.status_code, resp.text)
def _getJsonOrig(url): '''internal''' url = _URL_PREFIX + url resp = requests.get(urlparse(url).geturl(), proxies=_PYEX_PROXIES) if resp.status_code == 200: return resp.json() raise PyEXception('Response %d - ' % resp.status_code, resp.text)
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L98-L104
[ "def", "_getJsonOrig", "(", "url", ")", ":", "url", "=", "_URL_PREFIX", "+", "url", "resp", "=", "requests", ".", "get", "(", "urlparse", "(", "url", ")", ".", "geturl", "(", ")", ",", "proxies", "=", "_PYEX_PROXIES", ")", "if", "resp", ".", "status_code", "==", "200", ":", "return", "resp", ".", "json", "(", ")", "raise", "PyEXception", "(", "'Response %d - '", "%", "resp", ".", "status_code", ",", "resp", ".", "text", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_getJsonIEXCloud
for iex cloud
pyEX/common.py
def _getJsonIEXCloud(url, token='', version='beta'): '''for iex cloud''' url = _URL_PREFIX2.format(version=version) + url resp = requests.get(urlparse(url).geturl(), proxies=_PYEX_PROXIES, params={'token': token}) if resp.status_code == 200: return resp.json() raise PyEXception('Response %d - ' % resp.status_code, resp.text)
def _getJsonIEXCloud(url, token='', version='beta'): '''for iex cloud''' url = _URL_PREFIX2.format(version=version) + url resp = requests.get(urlparse(url).geturl(), proxies=_PYEX_PROXIES, params={'token': token}) if resp.status_code == 200: return resp.json() raise PyEXception('Response %d - ' % resp.status_code, resp.text)
[ "for", "iex", "cloud" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L107-L113
[ "def", "_getJsonIEXCloud", "(", "url", ",", "token", "=", "''", ",", "version", "=", "'beta'", ")", ":", "url", "=", "_URL_PREFIX2", ".", "format", "(", "version", "=", "version", ")", "+", "url", "resp", "=", "requests", ".", "get", "(", "urlparse", "(", "url", ")", ".", "geturl", "(", ")", ",", "proxies", "=", "_PYEX_PROXIES", ",", "params", "=", "{", "'token'", ":", "token", "}", ")", "if", "resp", ".", "status_code", "==", "200", ":", "return", "resp", ".", "json", "(", ")", "raise", "PyEXception", "(", "'Response %d - '", "%", "resp", ".", "status_code", ",", "resp", ".", "text", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_strOrDate
internal
pyEX/common.py
def _strOrDate(st): '''internal''' if isinstance(st, string_types): return st elif isinstance(st, datetime): return st.strftime('%Y%m%d') raise PyEXception('Not a date: %s', str(st))
def _strOrDate(st): '''internal''' if isinstance(st, string_types): return st elif isinstance(st, datetime): return st.strftime('%Y%m%d') raise PyEXception('Not a date: %s', str(st))
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L133-L139
[ "def", "_strOrDate", "(", "st", ")", ":", "if", "isinstance", "(", "st", ",", "string_types", ")", ":", "return", "st", "elif", "isinstance", "(", "st", ",", "datetime", ")", ":", "return", "st", ".", "strftime", "(", "'%Y%m%d'", ")", "raise", "PyEXception", "(", "'Not a date: %s'", ",", "str", "(", "st", ")", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_raiseIfNotStr
internal
pyEX/common.py
def _raiseIfNotStr(s): '''internal''' if s is not None and not isinstance(s, string_types): raise PyEXception('Cannot use type %s' % str(type(s)))
def _raiseIfNotStr(s): '''internal''' if s is not None and not isinstance(s, string_types): raise PyEXception('Cannot use type %s' % str(type(s)))
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L142-L145
[ "def", "_raiseIfNotStr", "(", "s", ")", ":", "if", "s", "is", "not", "None", "and", "not", "isinstance", "(", "s", ",", "string_types", ")", ":", "raise", "PyEXception", "(", "'Cannot use type %s'", "%", "str", "(", "type", "(", "s", ")", ")", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_tryJson
internal
pyEX/common.py
def _tryJson(data, raw=True): '''internal''' if raw: return data try: return json.loads(data) except ValueError: return data
def _tryJson(data, raw=True): '''internal''' if raw: return data try: return json.loads(data) except ValueError: return data
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L148-L155
[ "def", "_tryJson", "(", "data", ",", "raw", "=", "True", ")", ":", "if", "raw", ":", "return", "data", "try", ":", "return", "json", ".", "loads", "(", "data", ")", "except", "ValueError", ":", "return", "data" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_stream
internal
pyEX/common.py
def _stream(url, sendinit=None, on_data=print): '''internal''' cl = WSClient(url, sendinit=sendinit, on_data=on_data) return cl
def _stream(url, sendinit=None, on_data=print): '''internal''' cl = WSClient(url, sendinit=sendinit, on_data=on_data) return cl
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L192-L195
[ "def", "_stream", "(", "url", ",", "sendinit", "=", "None", ",", "on_data", "=", "print", ")", ":", "cl", "=", "WSClient", "(", "url", ",", "sendinit", "=", "sendinit", ",", "on_data", "=", "on_data", ")", "return", "cl" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_streamSSE
internal
pyEX/common.py
def _streamSSE(url, on_data=print, accrue=False): '''internal''' messages = SSEClient(url) if accrue: ret = [] for msg in messages: data = msg.data on_data(json.loads(data)) if accrue: ret.append(msg) return ret
def _streamSSE(url, on_data=print, accrue=False): '''internal''' messages = SSEClient(url) if accrue: ret = [] for msg in messages: data = msg.data on_data(json.loads(data)) if accrue: ret.append(msg) return ret
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L198-L210
[ "def", "_streamSSE", "(", "url", ",", "on_data", "=", "print", ",", "accrue", "=", "False", ")", ":", "messages", "=", "SSEClient", "(", "url", ")", "if", "accrue", ":", "ret", "=", "[", "]", "for", "msg", "in", "messages", ":", "data", "=", "msg", ".", "data", "on_data", "(", "json", ".", "loads", "(", "data", ")", ")", "if", "accrue", ":", "ret", ".", "append", "(", "msg", ")", "return", "ret" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_reindex
internal
pyEX/common.py
def _reindex(df, col): '''internal''' if col in df.columns: df.set_index(col, inplace=True)
def _reindex(df, col): '''internal''' if col in df.columns: df.set_index(col, inplace=True)
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L213-L216
[ "def", "_reindex", "(", "df", ",", "col", ")", ":", "if", "col", "in", "df", ".", "columns", ":", "df", ".", "set_index", "(", "col", ",", "inplace", "=", "True", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
_toDatetime
internal
pyEX/common.py
def _toDatetime(df, cols=None, tcols=None): '''internal''' cols = cols or _STANDARD_DATE_FIELDS tcols = tcols = _STANDARD_TIME_FIELDS for col in cols: if col in df: df[col] = pd.to_datetime(df[col], infer_datetime_format=True, errors='coerce') for tcol in tcols: if tcol in df: df[tcol] = pd.to_datetime(df[tcol], unit='ms', errors='coerce')
def _toDatetime(df, cols=None, tcols=None): '''internal''' cols = cols or _STANDARD_DATE_FIELDS tcols = tcols = _STANDARD_TIME_FIELDS for col in cols: if col in df: df[col] = pd.to_datetime(df[col], infer_datetime_format=True, errors='coerce') for tcol in tcols: if tcol in df: df[tcol] = pd.to_datetime(df[tcol], unit='ms', errors='coerce')
[ "internal" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/common.py#L219-L230
[ "def", "_toDatetime", "(", "df", ",", "cols", "=", "None", ",", "tcols", "=", "None", ")", ":", "cols", "=", "cols", "or", "_STANDARD_DATE_FIELDS", "tcols", "=", "tcols", "=", "_STANDARD_TIME_FIELDS", "for", "col", "in", "cols", ":", "if", "col", "in", "df", ":", "df", "[", "col", "]", "=", "pd", ".", "to_datetime", "(", "df", "[", "col", "]", ",", "infer_datetime_format", "=", "True", ",", "errors", "=", "'coerce'", ")", "for", "tcol", "in", "tcols", ":", "if", "tcol", "in", "df", ":", "df", "[", "tcol", "]", "=", "pd", ".", "to_datetime", "(", "df", "[", "tcol", "]", ",", "unit", "=", "'ms'", ",", "errors", "=", "'coerce'", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
balanceSheet
Pulls balance sheet data. Available quarterly (4 quarters) and annually (4 years) https://iexcloud.io/docs/api/#balance-sheet Updates at 8am, 9am UTC daily Args: symbol (string); Ticker to request token (string); Access token version (string); API version Returns: dict: result
pyEX/stocks.py
def balanceSheet(symbol, token='', version=''): '''Pulls balance sheet data. Available quarterly (4 quarters) and annually (4 years) https://iexcloud.io/docs/api/#balance-sheet Updates at 8am, 9am UTC daily Args: symbol (string); Ticker to request token (string); Access token version (string); API version Returns: dict: result ''' _raiseIfNotStr(symbol) return _getJson('stock/' + symbol + '/balance-sheet', token, version)
def balanceSheet(symbol, token='', version=''): '''Pulls balance sheet data. Available quarterly (4 quarters) and annually (4 years) https://iexcloud.io/docs/api/#balance-sheet Updates at 8am, 9am UTC daily Args: symbol (string); Ticker to request token (string); Access token version (string); API version Returns: dict: result ''' _raiseIfNotStr(symbol) return _getJson('stock/' + symbol + '/balance-sheet', token, version)
[ "Pulls", "balance", "sheet", "data", ".", "Available", "quarterly", "(", "4", "quarters", ")", "and", "annually", "(", "4", "years", ")" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/stocks.py#L12-L28
[ "def", "balanceSheet", "(", "symbol", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "_raiseIfNotStr", "(", "symbol", ")", "return", "_getJson", "(", "'stock/'", "+", "symbol", "+", "'/balance-sheet'", ",", "token", ",", "version", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
balanceSheetDF
Pulls balance sheet data. Available quarterly (4 quarters) and annually (4 years) https://iexcloud.io/docs/api/#balance-sheet Updates at 8am, 9am UTC daily Args: symbol (string); Ticker to request token (string); Access token version (string); API version Returns: DataFrame: result
pyEX/stocks.py
def balanceSheetDF(symbol, token='', version=''): '''Pulls balance sheet data. Available quarterly (4 quarters) and annually (4 years) https://iexcloud.io/docs/api/#balance-sheet Updates at 8am, 9am UTC daily Args: symbol (string); Ticker to request token (string); Access token version (string); API version Returns: DataFrame: result ''' val = balanceSheet(symbol, token, version) df = pd.io.json.json_normalize(val, 'balancesheet', 'symbol') _toDatetime(df) _reindex(df, 'reportDate') return df
def balanceSheetDF(symbol, token='', version=''): '''Pulls balance sheet data. Available quarterly (4 quarters) and annually (4 years) https://iexcloud.io/docs/api/#balance-sheet Updates at 8am, 9am UTC daily Args: symbol (string); Ticker to request token (string); Access token version (string); API version Returns: DataFrame: result ''' val = balanceSheet(symbol, token, version) df = pd.io.json.json_normalize(val, 'balancesheet', 'symbol') _toDatetime(df) _reindex(df, 'reportDate') return df
[ "Pulls", "balance", "sheet", "data", ".", "Available", "quarterly", "(", "4", "quarters", ")", "and", "annually", "(", "4", "years", ")" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/stocks.py#L31-L50
[ "def", "balanceSheetDF", "(", "symbol", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "val", "=", "balanceSheet", "(", "symbol", ",", "token", ",", "version", ")", "df", "=", "pd", ".", "io", ".", "json", ".", "json_normalize", "(", "val", ",", "'balancesheet'", ",", "'symbol'", ")", "_toDatetime", "(", "df", ")", "_reindex", "(", "df", ",", "'reportDate'", ")", "return", "df" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba
valid
batch
Batch several data requests into one invocation https://iexcloud.io/docs/api/#batch-requests Args: symbols (list); List of tickers to request fields (list); List of fields to request range_ (string); Date range for chart last (int); token (string); Access token version (string); API version Returns: dict: results in json
pyEX/stocks.py
def batch(symbols, fields=None, range_='1m', last=10, token='', version=''): '''Batch several data requests into one invocation https://iexcloud.io/docs/api/#batch-requests Args: symbols (list); List of tickers to request fields (list); List of fields to request range_ (string); Date range for chart last (int); token (string); Access token version (string); API version Returns: dict: results in json ''' fields = fields or _BATCH_TYPES[:10] # limit 10 if not isinstance(symbols, [].__class__): if not isinstance(symbols, str): raise PyEXception('batch expects string or list of strings for symbols argument') if isinstance(fields, str): fields = [fields] if range_ not in _TIMEFRAME_CHART: raise PyEXception('Range must be in %s' % str(_TIMEFRAME_CHART)) if isinstance(symbols, str): route = 'stock/{}/batch?types={}&range={}&last={}'.format(symbols, ','.join(fields), range_, last) return _getJson(route, token, version) if len(symbols) > 100: raise PyEXception('IEX will only handle up to 100 symbols at a time!') route = 'stock/market/batch?symbols={}&types={}&range={}&last={}'.format(','.join(symbols), ','.join(fields), range_, last) return _getJson(route, token, version)
def batch(symbols, fields=None, range_='1m', last=10, token='', version=''): '''Batch several data requests into one invocation https://iexcloud.io/docs/api/#batch-requests Args: symbols (list); List of tickers to request fields (list); List of fields to request range_ (string); Date range for chart last (int); token (string); Access token version (string); API version Returns: dict: results in json ''' fields = fields or _BATCH_TYPES[:10] # limit 10 if not isinstance(symbols, [].__class__): if not isinstance(symbols, str): raise PyEXception('batch expects string or list of strings for symbols argument') if isinstance(fields, str): fields = [fields] if range_ not in _TIMEFRAME_CHART: raise PyEXception('Range must be in %s' % str(_TIMEFRAME_CHART)) if isinstance(symbols, str): route = 'stock/{}/batch?types={}&range={}&last={}'.format(symbols, ','.join(fields), range_, last) return _getJson(route, token, version) if len(symbols) > 100: raise PyEXception('IEX will only handle up to 100 symbols at a time!') route = 'stock/market/batch?symbols={}&types={}&range={}&last={}'.format(','.join(symbols), ','.join(fields), range_, last) return _getJson(route, token, version)
[ "Batch", "several", "data", "requests", "into", "one", "invocation" ]
timkpaine/pyEX
python
https://github.com/timkpaine/pyEX/blob/91cf751dafdb208a0c8b5377945e5808b99f94ba/pyEX/stocks.py#L53-L89
[ "def", "batch", "(", "symbols", ",", "fields", "=", "None", ",", "range_", "=", "'1m'", ",", "last", "=", "10", ",", "token", "=", "''", ",", "version", "=", "''", ")", ":", "fields", "=", "fields", "or", "_BATCH_TYPES", "[", ":", "10", "]", "# limit 10", "if", "not", "isinstance", "(", "symbols", ",", "[", "]", ".", "__class__", ")", ":", "if", "not", "isinstance", "(", "symbols", ",", "str", ")", ":", "raise", "PyEXception", "(", "'batch expects string or list of strings for symbols argument'", ")", "if", "isinstance", "(", "fields", ",", "str", ")", ":", "fields", "=", "[", "fields", "]", "if", "range_", "not", "in", "_TIMEFRAME_CHART", ":", "raise", "PyEXception", "(", "'Range must be in %s'", "%", "str", "(", "_TIMEFRAME_CHART", ")", ")", "if", "isinstance", "(", "symbols", ",", "str", ")", ":", "route", "=", "'stock/{}/batch?types={}&range={}&last={}'", ".", "format", "(", "symbols", ",", "','", ".", "join", "(", "fields", ")", ",", "range_", ",", "last", ")", "return", "_getJson", "(", "route", ",", "token", ",", "version", ")", "if", "len", "(", "symbols", ")", ">", "100", ":", "raise", "PyEXception", "(", "'IEX will only handle up to 100 symbols at a time!'", ")", "route", "=", "'stock/market/batch?symbols={}&types={}&range={}&last={}'", ".", "format", "(", "','", ".", "join", "(", "symbols", ")", ",", "','", ".", "join", "(", "fields", ")", ",", "range_", ",", "last", ")", "return", "_getJson", "(", "route", ",", "token", ",", "version", ")" ]
91cf751dafdb208a0c8b5377945e5808b99f94ba