repository_name
stringlengths
7
55
func_path_in_repository
stringlengths
4
223
func_name
stringlengths
1
134
whole_func_string
stringlengths
75
104k
language
stringclasses
1 value
func_code_string
stringlengths
75
104k
func_code_tokens
listlengths
19
28.4k
func_documentation_string
stringlengths
1
46.9k
func_documentation_tokens
listlengths
1
1.97k
split_name
stringclasses
1 value
func_code_url
stringlengths
87
315
zetaops/zengine
zengine/engine.py
ZEngine.start_engine
def start_engine(self, **kwargs): """ Initializes the workflow with given request, response objects and diagram name. Args: session: input: workflow_name (str): Name of workflow diagram without ".bpmn" suffix. File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS` """ self.current = WFCurrent(**kwargs) self.wf_state = {'in_external': False, 'finished': False} if not self.current.new_token: self.wf_state = self.current.wf_cache.get(self.wf_state) self.current.workflow_name = self.wf_state['name'] # if we have a pre-selected object to work with, # inserting it as current.input['id'] and task_data['object_id'] if 'subject' in self.wf_state: self.current.input['id'] = self.wf_state['subject'] self.current.task_data['object_id'] = self.wf_state['subject'] self.check_for_authentication() self.check_for_permission() self.workflow = self.load_or_create_workflow() # if form data exists in input (user submitted) # put form data in wf task_data if 'form' in self.current.input: form = self.current.input['form'] if 'form_name' in form: self.current.task_data[form['form_name']] = form # in wf diagram, if property is stated as init = True # demanded initial values are assigned and put to cache start_init_values = self.workflow_spec.wf_properties.get('init', 'False') == 'True' if start_init_values: WFInit = get_object_from_path(settings.WF_INITIAL_VALUES)() WFInit.assign_wf_initial_values(self.current) log_msg = ("\n\n::::::::::: ENGINE STARTED :::::::::::\n" "\tWF: %s (Possible) TASK:%s\n" "\tCMD:%s\n" "\tSUBCMD:%s" % ( self.workflow.name, self.workflow.get_tasks(Task.READY), self.current.input.get('cmd'), self.current.input.get('subcmd'))) log.debug(log_msg) sys._zops_wf_state_log = log_msg self.current.workflow = self.workflow
python
def start_engine(self, **kwargs): """ Initializes the workflow with given request, response objects and diagram name. Args: session: input: workflow_name (str): Name of workflow diagram without ".bpmn" suffix. File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS` """ self.current = WFCurrent(**kwargs) self.wf_state = {'in_external': False, 'finished': False} if not self.current.new_token: self.wf_state = self.current.wf_cache.get(self.wf_state) self.current.workflow_name = self.wf_state['name'] # if we have a pre-selected object to work with, # inserting it as current.input['id'] and task_data['object_id'] if 'subject' in self.wf_state: self.current.input['id'] = self.wf_state['subject'] self.current.task_data['object_id'] = self.wf_state['subject'] self.check_for_authentication() self.check_for_permission() self.workflow = self.load_or_create_workflow() # if form data exists in input (user submitted) # put form data in wf task_data if 'form' in self.current.input: form = self.current.input['form'] if 'form_name' in form: self.current.task_data[form['form_name']] = form # in wf diagram, if property is stated as init = True # demanded initial values are assigned and put to cache start_init_values = self.workflow_spec.wf_properties.get('init', 'False') == 'True' if start_init_values: WFInit = get_object_from_path(settings.WF_INITIAL_VALUES)() WFInit.assign_wf_initial_values(self.current) log_msg = ("\n\n::::::::::: ENGINE STARTED :::::::::::\n" "\tWF: %s (Possible) TASK:%s\n" "\tCMD:%s\n" "\tSUBCMD:%s" % ( self.workflow.name, self.workflow.get_tasks(Task.READY), self.current.input.get('cmd'), self.current.input.get('subcmd'))) log.debug(log_msg) sys._zops_wf_state_log = log_msg self.current.workflow = self.workflow
[ "def", "start_engine", "(", "self", ",", "*", "*", "kwargs", ")", ":", "self", ".", "current", "=", "WFCurrent", "(", "*", "*", "kwargs", ")", "self", ".", "wf_state", "=", "{", "'in_external'", ":", "False", ",", "'finished'", ":", "False", "}", "if", "not", "self", ".", "current", ".", "new_token", ":", "self", ".", "wf_state", "=", "self", ".", "current", ".", "wf_cache", ".", "get", "(", "self", ".", "wf_state", ")", "self", ".", "current", ".", "workflow_name", "=", "self", ".", "wf_state", "[", "'name'", "]", "# if we have a pre-selected object to work with,", "# inserting it as current.input['id'] and task_data['object_id']", "if", "'subject'", "in", "self", ".", "wf_state", ":", "self", ".", "current", ".", "input", "[", "'id'", "]", "=", "self", ".", "wf_state", "[", "'subject'", "]", "self", ".", "current", ".", "task_data", "[", "'object_id'", "]", "=", "self", ".", "wf_state", "[", "'subject'", "]", "self", ".", "check_for_authentication", "(", ")", "self", ".", "check_for_permission", "(", ")", "self", ".", "workflow", "=", "self", ".", "load_or_create_workflow", "(", ")", "# if form data exists in input (user submitted)", "# put form data in wf task_data", "if", "'form'", "in", "self", ".", "current", ".", "input", ":", "form", "=", "self", ".", "current", ".", "input", "[", "'form'", "]", "if", "'form_name'", "in", "form", ":", "self", ".", "current", ".", "task_data", "[", "form", "[", "'form_name'", "]", "]", "=", "form", "# in wf diagram, if property is stated as init = True", "# demanded initial values are assigned and put to cache", "start_init_values", "=", "self", ".", "workflow_spec", ".", "wf_properties", ".", "get", "(", "'init'", ",", "'False'", ")", "==", "'True'", "if", "start_init_values", ":", "WFInit", "=", "get_object_from_path", "(", "settings", ".", "WF_INITIAL_VALUES", ")", "(", ")", "WFInit", ".", "assign_wf_initial_values", "(", "self", ".", "current", ")", "log_msg", "=", "(", "\"\\n\\n::::::::::: ENGINE STARTED :::::::::::\\n\"", "\"\\tWF: %s (Possible) TASK:%s\\n\"", "\"\\tCMD:%s\\n\"", "\"\\tSUBCMD:%s\"", "%", "(", "self", ".", "workflow", ".", "name", ",", "self", ".", "workflow", ".", "get_tasks", "(", "Task", ".", "READY", ")", ",", "self", ".", "current", ".", "input", ".", "get", "(", "'cmd'", ")", ",", "self", ".", "current", ".", "input", ".", "get", "(", "'subcmd'", ")", ")", ")", "log", ".", "debug", "(", "log_msg", ")", "sys", ".", "_zops_wf_state_log", "=", "log_msg", "self", ".", "current", ".", "workflow", "=", "self", ".", "workflow" ]
Initializes the workflow with given request, response objects and diagram name. Args: session: input: workflow_name (str): Name of workflow diagram without ".bpmn" suffix. File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
[ "Initializes", "the", "workflow", "with", "given", "request", "response", "objects", "and", "diagram", "name", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L241-L289
zetaops/zengine
zengine/engine.py
ZEngine.generate_wf_state_log
def generate_wf_state_log(self): """ Logs the state of workflow and content of task_data. """ output = '\n- - - - - -\n' output += "WORKFLOW: %s ( %s )" % (self.current.workflow_name.upper(), self.current.workflow.name) output += "\nTASK: %s ( %s )\n" % (self.current.task_name, self.current.task_type) output += "DATA:" for k, v in self.current.task_data.items(): if v: output += "\n\t%s: %s" % (k, v) output += "\nCURRENT:" output += "\n\tACTIVITY: %s" % self.current.activity output += "\n\tPOOL: %s" % self.current.pool output += "\n\tIN EXTERNAL: %s" % self.wf_state['in_external'] output += "\n\tLANE: %s" % self.current.lane_name output += "\n\tTOKEN: %s" % self.current.token sys._zops_wf_state_log = output return output
python
def generate_wf_state_log(self): """ Logs the state of workflow and content of task_data. """ output = '\n- - - - - -\n' output += "WORKFLOW: %s ( %s )" % (self.current.workflow_name.upper(), self.current.workflow.name) output += "\nTASK: %s ( %s )\n" % (self.current.task_name, self.current.task_type) output += "DATA:" for k, v in self.current.task_data.items(): if v: output += "\n\t%s: %s" % (k, v) output += "\nCURRENT:" output += "\n\tACTIVITY: %s" % self.current.activity output += "\n\tPOOL: %s" % self.current.pool output += "\n\tIN EXTERNAL: %s" % self.wf_state['in_external'] output += "\n\tLANE: %s" % self.current.lane_name output += "\n\tTOKEN: %s" % self.current.token sys._zops_wf_state_log = output return output
[ "def", "generate_wf_state_log", "(", "self", ")", ":", "output", "=", "'\\n- - - - - -\\n'", "output", "+=", "\"WORKFLOW: %s ( %s )\"", "%", "(", "self", ".", "current", ".", "workflow_name", ".", "upper", "(", ")", ",", "self", ".", "current", ".", "workflow", ".", "name", ")", "output", "+=", "\"\\nTASK: %s ( %s )\\n\"", "%", "(", "self", ".", "current", ".", "task_name", ",", "self", ".", "current", ".", "task_type", ")", "output", "+=", "\"DATA:\"", "for", "k", ",", "v", "in", "self", ".", "current", ".", "task_data", ".", "items", "(", ")", ":", "if", "v", ":", "output", "+=", "\"\\n\\t%s: %s\"", "%", "(", "k", ",", "v", ")", "output", "+=", "\"\\nCURRENT:\"", "output", "+=", "\"\\n\\tACTIVITY: %s\"", "%", "self", ".", "current", ".", "activity", "output", "+=", "\"\\n\\tPOOL: %s\"", "%", "self", ".", "current", ".", "pool", "output", "+=", "\"\\n\\tIN EXTERNAL: %s\"", "%", "self", ".", "wf_state", "[", "'in_external'", "]", "output", "+=", "\"\\n\\tLANE: %s\"", "%", "self", ".", "current", ".", "lane_name", "output", "+=", "\"\\n\\tTOKEN: %s\"", "%", "self", ".", "current", ".", "token", "sys", ".", "_zops_wf_state_log", "=", "output", "return", "output" ]
Logs the state of workflow and content of task_data.
[ "Logs", "the", "state", "of", "workflow", "and", "content", "of", "task_data", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L291-L311
zetaops/zengine
zengine/engine.py
ZEngine.switch_from_external_to_main_wf
def switch_from_external_to_main_wf(self): """ Main workflow switcher. This method recreates main workflow from `main wf` dict which was set by external workflow swicther previously. """ # in external assigned as True in switch_to_external_wf. # external_wf should finish EndEvent and it's name should be # also EndEvent for switching again to main wf. if self.wf_state['in_external'] and self.current.task_type == 'EndEvent' and \ self.current.task_name == 'EndEvent': # main_wf information was copied in switch_to_external_wf and it takes this information. main_wf = self.wf_state['main_wf'] # main_wf_name is assigned to current workflow name again. self.current.workflow_name = main_wf['name'] # For external WF, check permission and authentication. But after cleaning current task. self._clear_current_task() # check for auth and perm. current task cleared, do against new workflow_name self.check_for_authentication() self.check_for_permission() # WF knowledge is taken for main wf. self.workflow_spec = self.get_worfklow_spec() # WF instance is started again where leave off. self.workflow = self.deserialize_workflow(main_wf['step']) # Current WF is this WF instance. self.current.workflow = self.workflow # in_external is assigned as False self.wf_state['in_external'] = False # finished is assigned as False, because still in progress. self.wf_state['finished'] = False # pool info of main_wf is assigned. self.wf_state['pool'] = main_wf['pool'] self.current.pool = self.wf_state['pool'] # With main_wf is executed. self.run()
python
def switch_from_external_to_main_wf(self): """ Main workflow switcher. This method recreates main workflow from `main wf` dict which was set by external workflow swicther previously. """ # in external assigned as True in switch_to_external_wf. # external_wf should finish EndEvent and it's name should be # also EndEvent for switching again to main wf. if self.wf_state['in_external'] and self.current.task_type == 'EndEvent' and \ self.current.task_name == 'EndEvent': # main_wf information was copied in switch_to_external_wf and it takes this information. main_wf = self.wf_state['main_wf'] # main_wf_name is assigned to current workflow name again. self.current.workflow_name = main_wf['name'] # For external WF, check permission and authentication. But after cleaning current task. self._clear_current_task() # check for auth and perm. current task cleared, do against new workflow_name self.check_for_authentication() self.check_for_permission() # WF knowledge is taken for main wf. self.workflow_spec = self.get_worfklow_spec() # WF instance is started again where leave off. self.workflow = self.deserialize_workflow(main_wf['step']) # Current WF is this WF instance. self.current.workflow = self.workflow # in_external is assigned as False self.wf_state['in_external'] = False # finished is assigned as False, because still in progress. self.wf_state['finished'] = False # pool info of main_wf is assigned. self.wf_state['pool'] = main_wf['pool'] self.current.pool = self.wf_state['pool'] # With main_wf is executed. self.run()
[ "def", "switch_from_external_to_main_wf", "(", "self", ")", ":", "# in external assigned as True in switch_to_external_wf.", "# external_wf should finish EndEvent and it's name should be", "# also EndEvent for switching again to main wf.", "if", "self", ".", "wf_state", "[", "'in_external'", "]", "and", "self", ".", "current", ".", "task_type", "==", "'EndEvent'", "and", "self", ".", "current", ".", "task_name", "==", "'EndEvent'", ":", "# main_wf information was copied in switch_to_external_wf and it takes this information.", "main_wf", "=", "self", ".", "wf_state", "[", "'main_wf'", "]", "# main_wf_name is assigned to current workflow name again.", "self", ".", "current", ".", "workflow_name", "=", "main_wf", "[", "'name'", "]", "# For external WF, check permission and authentication. But after cleaning current task.", "self", ".", "_clear_current_task", "(", ")", "# check for auth and perm. current task cleared, do against new workflow_name", "self", ".", "check_for_authentication", "(", ")", "self", ".", "check_for_permission", "(", ")", "# WF knowledge is taken for main wf.", "self", ".", "workflow_spec", "=", "self", ".", "get_worfklow_spec", "(", ")", "# WF instance is started again where leave off.", "self", ".", "workflow", "=", "self", ".", "deserialize_workflow", "(", "main_wf", "[", "'step'", "]", ")", "# Current WF is this WF instance.", "self", ".", "current", ".", "workflow", "=", "self", ".", "workflow", "# in_external is assigned as False", "self", ".", "wf_state", "[", "'in_external'", "]", "=", "False", "# finished is assigned as False, because still in progress.", "self", ".", "wf_state", "[", "'finished'", "]", "=", "False", "# pool info of main_wf is assigned.", "self", ".", "wf_state", "[", "'pool'", "]", "=", "main_wf", "[", "'pool'", "]", "self", ".", "current", ".", "pool", "=", "self", ".", "wf_state", "[", "'pool'", "]", "# With main_wf is executed.", "self", ".", "run", "(", ")" ]
Main workflow switcher. This method recreates main workflow from `main wf` dict which was set by external workflow swicther previously.
[ "Main", "workflow", "switcher", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L316-L365
zetaops/zengine
zengine/engine.py
ZEngine.switch_to_external_wf
def switch_to_external_wf(self): """ External workflow switcher. This method copies main workflow information into a temporary dict `main_wf` and makes external workflow acting as main workflow. """ # External WF name should be stated at main wf diagram and type should be service task. if (self.current.task_type == 'ServiceTask' and self.current.task.task_spec.type == 'external'): log.debug("Entering to EXTERNAL WF") # Main wf information is copied to main_wf. main_wf = self.wf_state.copy() # workflow name from main wf diagram is assigned to current workflow name. # workflow name must be either in task_data with key 'external_wf' or in main diagram's # topic. self.current.workflow_name = self.current.task_data.pop('external_wf', False) or self.\ current.task.task_spec.topic # For external WF, check permission and authentication. But after cleaning current task. self._clear_current_task() # check for auth and perm. current task cleared, do against new workflow_name self.check_for_authentication() self.check_for_permission() # wf knowledge is taken for external wf. self.workflow_spec = self.get_worfklow_spec() # New WF instance is created for external wf. self.workflow = self.create_workflow() # Current WF is this WF instance. self.current.workflow = self.workflow # main_wf: main wf information. # in_external: it states external wf in progress. # finished: it shows that main wf didn't finish still progress in external wf. self.wf_state = {'main_wf': main_wf, 'in_external': True, 'finished': False}
python
def switch_to_external_wf(self): """ External workflow switcher. This method copies main workflow information into a temporary dict `main_wf` and makes external workflow acting as main workflow. """ # External WF name should be stated at main wf diagram and type should be service task. if (self.current.task_type == 'ServiceTask' and self.current.task.task_spec.type == 'external'): log.debug("Entering to EXTERNAL WF") # Main wf information is copied to main_wf. main_wf = self.wf_state.copy() # workflow name from main wf diagram is assigned to current workflow name. # workflow name must be either in task_data with key 'external_wf' or in main diagram's # topic. self.current.workflow_name = self.current.task_data.pop('external_wf', False) or self.\ current.task.task_spec.topic # For external WF, check permission and authentication. But after cleaning current task. self._clear_current_task() # check for auth and perm. current task cleared, do against new workflow_name self.check_for_authentication() self.check_for_permission() # wf knowledge is taken for external wf. self.workflow_spec = self.get_worfklow_spec() # New WF instance is created for external wf. self.workflow = self.create_workflow() # Current WF is this WF instance. self.current.workflow = self.workflow # main_wf: main wf information. # in_external: it states external wf in progress. # finished: it shows that main wf didn't finish still progress in external wf. self.wf_state = {'main_wf': main_wf, 'in_external': True, 'finished': False}
[ "def", "switch_to_external_wf", "(", "self", ")", ":", "# External WF name should be stated at main wf diagram and type should be service task.", "if", "(", "self", ".", "current", ".", "task_type", "==", "'ServiceTask'", "and", "self", ".", "current", ".", "task", ".", "task_spec", ".", "type", "==", "'external'", ")", ":", "log", ".", "debug", "(", "\"Entering to EXTERNAL WF\"", ")", "# Main wf information is copied to main_wf.", "main_wf", "=", "self", ".", "wf_state", ".", "copy", "(", ")", "# workflow name from main wf diagram is assigned to current workflow name.", "# workflow name must be either in task_data with key 'external_wf' or in main diagram's", "# topic.", "self", ".", "current", ".", "workflow_name", "=", "self", ".", "current", ".", "task_data", ".", "pop", "(", "'external_wf'", ",", "False", ")", "or", "self", ".", "current", ".", "task", ".", "task_spec", ".", "topic", "# For external WF, check permission and authentication. But after cleaning current task.", "self", ".", "_clear_current_task", "(", ")", "# check for auth and perm. current task cleared, do against new workflow_name", "self", ".", "check_for_authentication", "(", ")", "self", ".", "check_for_permission", "(", ")", "# wf knowledge is taken for external wf.", "self", ".", "workflow_spec", "=", "self", ".", "get_worfklow_spec", "(", ")", "# New WF instance is created for external wf.", "self", ".", "workflow", "=", "self", ".", "create_workflow", "(", ")", "# Current WF is this WF instance.", "self", ".", "current", ".", "workflow", "=", "self", ".", "workflow", "# main_wf: main wf information.", "# in_external: it states external wf in progress.", "# finished: it shows that main wf didn't finish still progress in external wf.", "self", ".", "wf_state", "=", "{", "'main_wf'", ":", "main_wf", ",", "'in_external'", ":", "True", ",", "'finished'", ":", "False", "}" ]
External workflow switcher. This method copies main workflow information into a temporary dict `main_wf` and makes external workflow acting as main workflow.
[ "External", "workflow", "switcher", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L367-L408
zetaops/zengine
zengine/engine.py
ZEngine._clear_current_task
def _clear_current_task(self): """ Clear tasks related attributes, checks permissions While switching WF to WF, authentication and permissions are checked for new WF. """ self.current.task_name = None self.current.task_type = None self.current.task = None
python
def _clear_current_task(self): """ Clear tasks related attributes, checks permissions While switching WF to WF, authentication and permissions are checked for new WF. """ self.current.task_name = None self.current.task_type = None self.current.task = None
[ "def", "_clear_current_task", "(", "self", ")", ":", "self", ".", "current", ".", "task_name", "=", "None", "self", ".", "current", ".", "task_type", "=", "None", "self", ".", "current", ".", "task", "=", "None" ]
Clear tasks related attributes, checks permissions While switching WF to WF, authentication and permissions are checked for new WF.
[ "Clear", "tasks", "related", "attributes", "checks", "permissions", "While", "switching", "WF", "to", "WF", "authentication", "and", "permissions", "are", "checked", "for", "new", "WF", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L410-L418
zetaops/zengine
zengine/engine.py
ZEngine.run
def run(self): """ Main loop of the workflow engine - Updates ::class:`~WFCurrent` object. - Checks for Permissions. - Activates all READY tasks. - Runs referenced activities (method calls). - Saves WF states. - Stops if current task is a UserTask or EndTask. - Deletes state object if we finish the WF. """ # FIXME: raise if first task after line change isn't a UserTask # FIXME: raise if last task of a workflow is a UserTask # actually this check should be done at parser is_lane_changed = False while self._should_we_run(): self.check_for_rerun_user_task() task = None for task in self.workflow.get_tasks(state=Task.READY): self.current.old_lane = self.current.lane_name self.current._update_task(task) if self.catch_lane_change(): return self.check_for_permission() self.check_for_lane_permission() self.log_wf_state() self.switch_lang() self.run_activity() self.parse_workflow_messages() self.workflow.complete_task_from_id(self.current.task.id) self._save_or_delete_workflow() self.switch_to_external_wf() if task is None: break self.switch_from_external_to_main_wf() self.current.output['token'] = self.current.token # look for incoming ready task(s) for task in self.workflow.get_tasks(state=Task.READY): self.current._update_task(task) self.catch_lane_change() self.handle_wf_finalization()
python
def run(self): """ Main loop of the workflow engine - Updates ::class:`~WFCurrent` object. - Checks for Permissions. - Activates all READY tasks. - Runs referenced activities (method calls). - Saves WF states. - Stops if current task is a UserTask or EndTask. - Deletes state object if we finish the WF. """ # FIXME: raise if first task after line change isn't a UserTask # FIXME: raise if last task of a workflow is a UserTask # actually this check should be done at parser is_lane_changed = False while self._should_we_run(): self.check_for_rerun_user_task() task = None for task in self.workflow.get_tasks(state=Task.READY): self.current.old_lane = self.current.lane_name self.current._update_task(task) if self.catch_lane_change(): return self.check_for_permission() self.check_for_lane_permission() self.log_wf_state() self.switch_lang() self.run_activity() self.parse_workflow_messages() self.workflow.complete_task_from_id(self.current.task.id) self._save_or_delete_workflow() self.switch_to_external_wf() if task is None: break self.switch_from_external_to_main_wf() self.current.output['token'] = self.current.token # look for incoming ready task(s) for task in self.workflow.get_tasks(state=Task.READY): self.current._update_task(task) self.catch_lane_change() self.handle_wf_finalization()
[ "def", "run", "(", "self", ")", ":", "# FIXME: raise if first task after line change isn't a UserTask", "# FIXME: raise if last task of a workflow is a UserTask", "# actually this check should be done at parser", "is_lane_changed", "=", "False", "while", "self", ".", "_should_we_run", "(", ")", ":", "self", ".", "check_for_rerun_user_task", "(", ")", "task", "=", "None", "for", "task", "in", "self", ".", "workflow", ".", "get_tasks", "(", "state", "=", "Task", ".", "READY", ")", ":", "self", ".", "current", ".", "old_lane", "=", "self", ".", "current", ".", "lane_name", "self", ".", "current", ".", "_update_task", "(", "task", ")", "if", "self", ".", "catch_lane_change", "(", ")", ":", "return", "self", ".", "check_for_permission", "(", ")", "self", ".", "check_for_lane_permission", "(", ")", "self", ".", "log_wf_state", "(", ")", "self", ".", "switch_lang", "(", ")", "self", ".", "run_activity", "(", ")", "self", ".", "parse_workflow_messages", "(", ")", "self", ".", "workflow", ".", "complete_task_from_id", "(", "self", ".", "current", ".", "task", ".", "id", ")", "self", ".", "_save_or_delete_workflow", "(", ")", "self", ".", "switch_to_external_wf", "(", ")", "if", "task", "is", "None", ":", "break", "self", ".", "switch_from_external_to_main_wf", "(", ")", "self", ".", "current", ".", "output", "[", "'token'", "]", "=", "self", ".", "current", ".", "token", "# look for incoming ready task(s)", "for", "task", "in", "self", ".", "workflow", ".", "get_tasks", "(", "state", "=", "Task", ".", "READY", ")", ":", "self", ".", "current", ".", "_update_task", "(", "task", ")", "self", ".", "catch_lane_change", "(", ")", "self", ".", "handle_wf_finalization", "(", ")" ]
Main loop of the workflow engine - Updates ::class:`~WFCurrent` object. - Checks for Permissions. - Activates all READY tasks. - Runs referenced activities (method calls). - Saves WF states. - Stops if current task is a UserTask or EndTask. - Deletes state object if we finish the WF.
[ "Main", "loop", "of", "the", "workflow", "engine" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L432-L477
zetaops/zengine
zengine/engine.py
ZEngine.check_for_rerun_user_task
def check_for_rerun_user_task(self): """ Checks that the user task needs to re-run. If necessary, current task and pre task's states are changed and re-run. If wf_meta not in data(there is no user interaction from pre-task) and last completed task type is user task and current step is not EndEvent and there is no lane change, this user task is rerun. """ data = self.current.input if 'wf_meta' in data: return current_task = self.workflow.get_tasks(Task.READY)[0] current_task_type = current_task.task_spec.__class__.__name__ pre_task = current_task.parent pre_task_type = pre_task.task_spec.__class__.__name__ if pre_task_type != 'UserTask': return if current_task_type == 'EndEvent': return pre_lane = pre_task.task_spec.lane current_lane = current_task.task_spec.lane if pre_lane == current_lane: pre_task._set_state(Task.READY) current_task._set_state(Task.MAYBE)
python
def check_for_rerun_user_task(self): """ Checks that the user task needs to re-run. If necessary, current task and pre task's states are changed and re-run. If wf_meta not in data(there is no user interaction from pre-task) and last completed task type is user task and current step is not EndEvent and there is no lane change, this user task is rerun. """ data = self.current.input if 'wf_meta' in data: return current_task = self.workflow.get_tasks(Task.READY)[0] current_task_type = current_task.task_spec.__class__.__name__ pre_task = current_task.parent pre_task_type = pre_task.task_spec.__class__.__name__ if pre_task_type != 'UserTask': return if current_task_type == 'EndEvent': return pre_lane = pre_task.task_spec.lane current_lane = current_task.task_spec.lane if pre_lane == current_lane: pre_task._set_state(Task.READY) current_task._set_state(Task.MAYBE)
[ "def", "check_for_rerun_user_task", "(", "self", ")", ":", "data", "=", "self", ".", "current", ".", "input", "if", "'wf_meta'", "in", "data", ":", "return", "current_task", "=", "self", ".", "workflow", ".", "get_tasks", "(", "Task", ".", "READY", ")", "[", "0", "]", "current_task_type", "=", "current_task", ".", "task_spec", ".", "__class__", ".", "__name__", "pre_task", "=", "current_task", ".", "parent", "pre_task_type", "=", "pre_task", ".", "task_spec", ".", "__class__", ".", "__name__", "if", "pre_task_type", "!=", "'UserTask'", ":", "return", "if", "current_task_type", "==", "'EndEvent'", ":", "return", "pre_lane", "=", "pre_task", ".", "task_spec", ".", "lane", "current_lane", "=", "current_task", ".", "task_spec", ".", "lane", "if", "pre_lane", "==", "current_lane", ":", "pre_task", ".", "_set_state", "(", "Task", ".", "READY", ")", "current_task", ".", "_set_state", "(", "Task", ".", "MAYBE", ")" ]
Checks that the user task needs to re-run. If necessary, current task and pre task's states are changed and re-run. If wf_meta not in data(there is no user interaction from pre-task) and last completed task type is user task and current step is not EndEvent and there is no lane change, this user task is rerun.
[ "Checks", "that", "the", "user", "task", "needs", "to", "re", "-", "run", ".", "If", "necessary", "current", "task", "and", "pre", "task", "s", "states", "are", "changed", "and", "re", "-", "run", ".", "If", "wf_meta", "not", "in", "data", "(", "there", "is", "no", "user", "interaction", "from", "pre", "-", "task", ")", "and", "last", "completed", "task", "type", "is", "user", "task", "and", "current", "step", "is", "not", "EndEvent", "and", "there", "is", "no", "lane", "change", "this", "user", "task", "is", "rerun", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L479-L506
zetaops/zengine
zengine/engine.py
ZEngine.switch_lang
def switch_lang(self): """Switch to the language of the current user. If the current language is already the specified one, nothing will be done. """ locale = self.current.locale translation.InstalledLocale.install_language(locale['locale_language']) translation.InstalledLocale.install_locale(locale['locale_datetime'], 'datetime') translation.InstalledLocale.install_locale(locale['locale_number'], 'number')
python
def switch_lang(self): """Switch to the language of the current user. If the current language is already the specified one, nothing will be done. """ locale = self.current.locale translation.InstalledLocale.install_language(locale['locale_language']) translation.InstalledLocale.install_locale(locale['locale_datetime'], 'datetime') translation.InstalledLocale.install_locale(locale['locale_number'], 'number')
[ "def", "switch_lang", "(", "self", ")", ":", "locale", "=", "self", ".", "current", ".", "locale", "translation", ".", "InstalledLocale", ".", "install_language", "(", "locale", "[", "'locale_language'", "]", ")", "translation", ".", "InstalledLocale", ".", "install_locale", "(", "locale", "[", "'locale_datetime'", "]", ",", "'datetime'", ")", "translation", ".", "InstalledLocale", ".", "install_locale", "(", "locale", "[", "'locale_number'", "]", ",", "'number'", ")" ]
Switch to the language of the current user. If the current language is already the specified one, nothing will be done.
[ "Switch", "to", "the", "language", "of", "the", "current", "user", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L508-L516
zetaops/zengine
zengine/engine.py
ZEngine.catch_lane_change
def catch_lane_change(self): """ trigger a lane_user_change signal if we switched to a new lane and new lane's user is different from current one """ if self.current.lane_name: if self.current.old_lane and self.current.lane_name != self.current.old_lane: # if lane_name not found in pool or it's user different from the current(old) user if (self.current.lane_id not in self.current.pool or self.current.pool[self.current.lane_id] != self.current.user_id): self.current.log.info("LANE CHANGE : %s >> %s" % (self.current.old_lane, self.current.lane_name)) if self.current.lane_auto_sendoff: self.current.sendoff_current_user() self.current.flow_enabled = False if self.current.lane_auto_invite: self.current.invite_other_parties(self._get_possible_lane_owners()) return True
python
def catch_lane_change(self): """ trigger a lane_user_change signal if we switched to a new lane and new lane's user is different from current one """ if self.current.lane_name: if self.current.old_lane and self.current.lane_name != self.current.old_lane: # if lane_name not found in pool or it's user different from the current(old) user if (self.current.lane_id not in self.current.pool or self.current.pool[self.current.lane_id] != self.current.user_id): self.current.log.info("LANE CHANGE : %s >> %s" % (self.current.old_lane, self.current.lane_name)) if self.current.lane_auto_sendoff: self.current.sendoff_current_user() self.current.flow_enabled = False if self.current.lane_auto_invite: self.current.invite_other_parties(self._get_possible_lane_owners()) return True
[ "def", "catch_lane_change", "(", "self", ")", ":", "if", "self", ".", "current", ".", "lane_name", ":", "if", "self", ".", "current", ".", "old_lane", "and", "self", ".", "current", ".", "lane_name", "!=", "self", ".", "current", ".", "old_lane", ":", "# if lane_name not found in pool or it's user different from the current(old) user", "if", "(", "self", ".", "current", ".", "lane_id", "not", "in", "self", ".", "current", ".", "pool", "or", "self", ".", "current", ".", "pool", "[", "self", ".", "current", ".", "lane_id", "]", "!=", "self", ".", "current", ".", "user_id", ")", ":", "self", ".", "current", ".", "log", ".", "info", "(", "\"LANE CHANGE : %s >> %s\"", "%", "(", "self", ".", "current", ".", "old_lane", ",", "self", ".", "current", ".", "lane_name", ")", ")", "if", "self", ".", "current", ".", "lane_auto_sendoff", ":", "self", ".", "current", ".", "sendoff_current_user", "(", ")", "self", ".", "current", ".", "flow_enabled", "=", "False", "if", "self", ".", "current", ".", "lane_auto_invite", ":", "self", ".", "current", ".", "invite_other_parties", "(", "self", ".", "_get_possible_lane_owners", "(", ")", ")", "return", "True" ]
trigger a lane_user_change signal if we switched to a new lane and new lane's user is different from current one
[ "trigger", "a", "lane_user_change", "signal", "if", "we", "switched", "to", "a", "new", "lane", "and", "new", "lane", "s", "user", "is", "different", "from", "current", "one" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L518-L535
zetaops/zengine
zengine/engine.py
ZEngine.parse_workflow_messages
def parse_workflow_messages(self): """ Transmits client message that defined in a workflow task's inputOutput extension .. code-block:: xml <bpmn2:extensionElements> <camunda:inputOutput> <camunda:inputParameter name="client_message"> <camunda:map> <camunda:entry key="title">Teşekkürler</camunda:entry> <camunda:entry key="body">İşlem Başarılı</camunda:entry> <camunda:entry key="type">info</camunda:entry> </camunda:map> </camunda:inputParameter> </camunda:inputOutput> </bpmn2:extensionElements> """ if 'client_message' in self.current.spec.data: m = self.current.spec.data['client_message'] self.current.msg_box(title=m.get('title'), msg=m.get('body'), typ=m.get('type', 'info'))
python
def parse_workflow_messages(self): """ Transmits client message that defined in a workflow task's inputOutput extension .. code-block:: xml <bpmn2:extensionElements> <camunda:inputOutput> <camunda:inputParameter name="client_message"> <camunda:map> <camunda:entry key="title">Teşekkürler</camunda:entry> <camunda:entry key="body">İşlem Başarılı</camunda:entry> <camunda:entry key="type">info</camunda:entry> </camunda:map> </camunda:inputParameter> </camunda:inputOutput> </bpmn2:extensionElements> """ if 'client_message' in self.current.spec.data: m = self.current.spec.data['client_message'] self.current.msg_box(title=m.get('title'), msg=m.get('body'), typ=m.get('type', 'info'))
[ "def", "parse_workflow_messages", "(", "self", ")", ":", "if", "'client_message'", "in", "self", ".", "current", ".", "spec", ".", "data", ":", "m", "=", "self", ".", "current", ".", "spec", ".", "data", "[", "'client_message'", "]", "self", ".", "current", ".", "msg_box", "(", "title", "=", "m", ".", "get", "(", "'title'", ")", ",", "msg", "=", "m", ".", "get", "(", "'body'", ")", ",", "typ", "=", "m", ".", "get", "(", "'type'", ",", "'info'", ")", ")" ]
Transmits client message that defined in a workflow task's inputOutput extension .. code-block:: xml <bpmn2:extensionElements> <camunda:inputOutput> <camunda:inputParameter name="client_message"> <camunda:map> <camunda:entry key="title">Teşekkürler</camunda:entry> <camunda:entry key="body">İşlem Başarılı</camunda:entry> <camunda:entry key="type">info</camunda:entry> </camunda:map> </camunda:inputParameter> </camunda:inputOutput> </bpmn2:extensionElements>
[ "Transmits", "client", "message", "that", "defined", "in", "a", "workflow", "task", "s", "inputOutput", "extension" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L538-L562
zetaops/zengine
zengine/engine.py
ZEngine.run_activity
def run_activity(self): """ runs the method that referenced from current task """ activity = self.current.activity if activity: if activity not in self.wf_activities: self._load_activity(activity) self.current.log.debug( "Calling Activity %s from %s" % (activity, self.wf_activities[activity])) self.wf_activities[self.current.activity](self.current)
python
def run_activity(self): """ runs the method that referenced from current task """ activity = self.current.activity if activity: if activity not in self.wf_activities: self._load_activity(activity) self.current.log.debug( "Calling Activity %s from %s" % (activity, self.wf_activities[activity])) self.wf_activities[self.current.activity](self.current)
[ "def", "run_activity", "(", "self", ")", ":", "activity", "=", "self", ".", "current", ".", "activity", "if", "activity", ":", "if", "activity", "not", "in", "self", ".", "wf_activities", ":", "self", ".", "_load_activity", "(", "activity", ")", "self", ".", "current", ".", "log", ".", "debug", "(", "\"Calling Activity %s from %s\"", "%", "(", "activity", ",", "self", ".", "wf_activities", "[", "activity", "]", ")", ")", "self", ".", "wf_activities", "[", "self", ".", "current", ".", "activity", "]", "(", "self", ".", "current", ")" ]
runs the method that referenced from current task
[ "runs", "the", "method", "that", "referenced", "from", "current", "task" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L573-L583
zetaops/zengine
zengine/engine.py
ZEngine._import_object
def _import_object(self, path, look_for_cls_method): """ Imports the module that contains the referenced method. Args: path: python path of class/function look_for_cls_method (bool): If True, treat the last part of path as class method. Returns: Tuple. (class object, class name, method to be called) """ last_nth = 2 if look_for_cls_method else 1 path = path.split('.') module_path = '.'.join(path[:-last_nth]) class_name = path[-last_nth] module = importlib.import_module(module_path) if look_for_cls_method and path[-last_nth:][0] == path[-last_nth]: class_method = path[-last_nth:][1] else: class_method = None return getattr(module, class_name), class_name, class_method
python
def _import_object(self, path, look_for_cls_method): """ Imports the module that contains the referenced method. Args: path: python path of class/function look_for_cls_method (bool): If True, treat the last part of path as class method. Returns: Tuple. (class object, class name, method to be called) """ last_nth = 2 if look_for_cls_method else 1 path = path.split('.') module_path = '.'.join(path[:-last_nth]) class_name = path[-last_nth] module = importlib.import_module(module_path) if look_for_cls_method and path[-last_nth:][0] == path[-last_nth]: class_method = path[-last_nth:][1] else: class_method = None return getattr(module, class_name), class_name, class_method
[ "def", "_import_object", "(", "self", ",", "path", ",", "look_for_cls_method", ")", ":", "last_nth", "=", "2", "if", "look_for_cls_method", "else", "1", "path", "=", "path", ".", "split", "(", "'.'", ")", "module_path", "=", "'.'", ".", "join", "(", "path", "[", ":", "-", "last_nth", "]", ")", "class_name", "=", "path", "[", "-", "last_nth", "]", "module", "=", "importlib", ".", "import_module", "(", "module_path", ")", "if", "look_for_cls_method", "and", "path", "[", "-", "last_nth", ":", "]", "[", "0", "]", "==", "path", "[", "-", "last_nth", "]", ":", "class_method", "=", "path", "[", "-", "last_nth", ":", "]", "[", "1", "]", "else", ":", "class_method", "=", "None", "return", "getattr", "(", "module", ",", "class_name", ")", ",", "class_name", ",", "class_method" ]
Imports the module that contains the referenced method. Args: path: python path of class/function look_for_cls_method (bool): If True, treat the last part of path as class method. Returns: Tuple. (class object, class name, method to be called)
[ "Imports", "the", "module", "that", "contains", "the", "referenced", "method", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L585-L606
zetaops/zengine
zengine/engine.py
ZEngine._load_activity
def _load_activity(self, activity): """ Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path. """ fpths = [] full_path = '' errors = [] paths = settings.ACTIVITY_MODULES_IMPORT_PATHS number_of_paths = len(paths) for index_no in range(number_of_paths): full_path = "%s.%s" % (paths[index_no], activity) for look4kls in (0, 1): try: self.current.log.info("try to load from %s[%s]" % (full_path, look4kls)) kls, cls_name, cls_method = self._import_object(full_path, look4kls) if cls_method: self.current.log.info("WILLCall %s(current).%s()" % (kls, cls_method)) self.wf_activities[activity] = lambda crnt: getattr(kls(crnt), cls_method)() else: self.wf_activities[activity] = kls return except (ImportError, AttributeError): fpths.append(full_path) errmsg = "{activity} not found under these paths:\n\n >>> {paths} \n\n" \ "Error Messages:\n {errors}" errors.append("\n========================================================>\n" "| PATH | %s" "\n========================================================>\n\n" "%s" % (full_path, traceback.format_exc())) assert index_no != number_of_paths - 1, errmsg.format(activity=activity, paths='\n >>> '.join( set(fpths)), errors='\n\n'.join(errors) ) except: self.current.log.exception("Cannot found the %s" % activity)
python
def _load_activity(self, activity): """ Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path. """ fpths = [] full_path = '' errors = [] paths = settings.ACTIVITY_MODULES_IMPORT_PATHS number_of_paths = len(paths) for index_no in range(number_of_paths): full_path = "%s.%s" % (paths[index_no], activity) for look4kls in (0, 1): try: self.current.log.info("try to load from %s[%s]" % (full_path, look4kls)) kls, cls_name, cls_method = self._import_object(full_path, look4kls) if cls_method: self.current.log.info("WILLCall %s(current).%s()" % (kls, cls_method)) self.wf_activities[activity] = lambda crnt: getattr(kls(crnt), cls_method)() else: self.wf_activities[activity] = kls return except (ImportError, AttributeError): fpths.append(full_path) errmsg = "{activity} not found under these paths:\n\n >>> {paths} \n\n" \ "Error Messages:\n {errors}" errors.append("\n========================================================>\n" "| PATH | %s" "\n========================================================>\n\n" "%s" % (full_path, traceback.format_exc())) assert index_no != number_of_paths - 1, errmsg.format(activity=activity, paths='\n >>> '.join( set(fpths)), errors='\n\n'.join(errors) ) except: self.current.log.exception("Cannot found the %s" % activity)
[ "def", "_load_activity", "(", "self", ",", "activity", ")", ":", "fpths", "=", "[", "]", "full_path", "=", "''", "errors", "=", "[", "]", "paths", "=", "settings", ".", "ACTIVITY_MODULES_IMPORT_PATHS", "number_of_paths", "=", "len", "(", "paths", ")", "for", "index_no", "in", "range", "(", "number_of_paths", ")", ":", "full_path", "=", "\"%s.%s\"", "%", "(", "paths", "[", "index_no", "]", ",", "activity", ")", "for", "look4kls", "in", "(", "0", ",", "1", ")", ":", "try", ":", "self", ".", "current", ".", "log", ".", "info", "(", "\"try to load from %s[%s]\"", "%", "(", "full_path", ",", "look4kls", ")", ")", "kls", ",", "cls_name", ",", "cls_method", "=", "self", ".", "_import_object", "(", "full_path", ",", "look4kls", ")", "if", "cls_method", ":", "self", ".", "current", ".", "log", ".", "info", "(", "\"WILLCall %s(current).%s()\"", "%", "(", "kls", ",", "cls_method", ")", ")", "self", ".", "wf_activities", "[", "activity", "]", "=", "lambda", "crnt", ":", "getattr", "(", "kls", "(", "crnt", ")", ",", "cls_method", ")", "(", ")", "else", ":", "self", ".", "wf_activities", "[", "activity", "]", "=", "kls", "return", "except", "(", "ImportError", ",", "AttributeError", ")", ":", "fpths", ".", "append", "(", "full_path", ")", "errmsg", "=", "\"{activity} not found under these paths:\\n\\n >>> {paths} \\n\\n\"", "\"Error Messages:\\n {errors}\"", "errors", ".", "append", "(", "\"\\n========================================================>\\n\"", "\"| PATH | %s\"", "\"\\n========================================================>\\n\\n\"", "\"%s\"", "%", "(", "full_path", ",", "traceback", ".", "format_exc", "(", ")", ")", ")", "assert", "index_no", "!=", "number_of_paths", "-", "1", ",", "errmsg", ".", "format", "(", "activity", "=", "activity", ",", "paths", "=", "'\\n >>> '", ".", "join", "(", "set", "(", "fpths", ")", ")", ",", "errors", "=", "'\\n\\n'", ".", "join", "(", "errors", ")", ")", "except", ":", "self", ".", "current", ".", "log", ".", "exception", "(", "\"Cannot found the %s\"", "%", "activity", ")" ]
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
[ "Iterates", "trough", "the", "all", "enabled", "~zengine", ".", "settings", ".", "ACTIVITY_MODULES_IMPORT_PATHS", "to", "find", "the", "given", "path", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L608-L643
zetaops/zengine
zengine/engine.py
ZEngine.check_for_authentication
def check_for_authentication(self): """ Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list. Raises: HTTPUnauthorized: if WF needs an authenticated user and current user isn't. """ auth_required = self.current.workflow_name not in settings.ANONYMOUS_WORKFLOWS if auth_required and not self.current.is_auth: self.current.log.debug("LOGIN REQUIRED:::: %s" % self.current.workflow_name) raise HTTPError(401, "Login required for %s" % self.current.workflow_name)
python
def check_for_authentication(self): """ Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list. Raises: HTTPUnauthorized: if WF needs an authenticated user and current user isn't. """ auth_required = self.current.workflow_name not in settings.ANONYMOUS_WORKFLOWS if auth_required and not self.current.is_auth: self.current.log.debug("LOGIN REQUIRED:::: %s" % self.current.workflow_name) raise HTTPError(401, "Login required for %s" % self.current.workflow_name)
[ "def", "check_for_authentication", "(", "self", ")", ":", "auth_required", "=", "self", ".", "current", ".", "workflow_name", "not", "in", "settings", ".", "ANONYMOUS_WORKFLOWS", "if", "auth_required", "and", "not", "self", ".", "current", ".", "is_auth", ":", "self", ".", "current", ".", "log", ".", "debug", "(", "\"LOGIN REQUIRED:::: %s\"", "%", "self", ".", "current", ".", "workflow_name", ")", "raise", "HTTPError", "(", "401", ",", "\"Login required for %s\"", "%", "self", ".", "current", ".", "workflow_name", ")" ]
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list. Raises: HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
[ "Checks", "current", "workflow", "against", ":", "py", ":", "data", ":", "~zengine", ".", "settings", ".", "ANONYMOUS_WORKFLOWS", "list", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L645-L655
zetaops/zengine
zengine/engine.py
ZEngine.check_for_lane_permission
def check_for_lane_permission(self): """ One or more permissions can be associated with a lane of a workflow. In a similar way, a lane can be restricted with relation to other lanes of the workflow. This method called on lane changes and checks user has required permissions and relations. Raises: HTTPForbidden: if the current user hasn't got the required permissions and proper relations """ # TODO: Cache lane_data in app memory if self.current.lane_permission: log.debug("HAS LANE PERM: %s" % self.current.lane_permission) perm = self.current.lane_permission if not self.current.has_permission(perm): raise HTTPError(403, "You don't have required lane permission: %s" % perm) if self.current.lane_relations: context = self.get_pool_context() log.debug("HAS LANE RELS: %s" % self.current.lane_relations) try: cond_result = eval(self.current.lane_relations, context) except: log.exception("CONDITION EVAL ERROR : %s || %s" % ( self.current.lane_relations, context)) raise if not cond_result: log.debug("LANE RELATION ERR: %s %s" % (self.current.lane_relations, context)) raise HTTPError(403, "You aren't qualified for this lane: %s" % self.current.lane_relations)
python
def check_for_lane_permission(self): """ One or more permissions can be associated with a lane of a workflow. In a similar way, a lane can be restricted with relation to other lanes of the workflow. This method called on lane changes and checks user has required permissions and relations. Raises: HTTPForbidden: if the current user hasn't got the required permissions and proper relations """ # TODO: Cache lane_data in app memory if self.current.lane_permission: log.debug("HAS LANE PERM: %s" % self.current.lane_permission) perm = self.current.lane_permission if not self.current.has_permission(perm): raise HTTPError(403, "You don't have required lane permission: %s" % perm) if self.current.lane_relations: context = self.get_pool_context() log.debug("HAS LANE RELS: %s" % self.current.lane_relations) try: cond_result = eval(self.current.lane_relations, context) except: log.exception("CONDITION EVAL ERROR : %s || %s" % ( self.current.lane_relations, context)) raise if not cond_result: log.debug("LANE RELATION ERR: %s %s" % (self.current.lane_relations, context)) raise HTTPError(403, "You aren't qualified for this lane: %s" % self.current.lane_relations)
[ "def", "check_for_lane_permission", "(", "self", ")", ":", "# TODO: Cache lane_data in app memory", "if", "self", ".", "current", ".", "lane_permission", ":", "log", ".", "debug", "(", "\"HAS LANE PERM: %s\"", "%", "self", ".", "current", ".", "lane_permission", ")", "perm", "=", "self", ".", "current", ".", "lane_permission", "if", "not", "self", ".", "current", ".", "has_permission", "(", "perm", ")", ":", "raise", "HTTPError", "(", "403", ",", "\"You don't have required lane permission: %s\"", "%", "perm", ")", "if", "self", ".", "current", ".", "lane_relations", ":", "context", "=", "self", ".", "get_pool_context", "(", ")", "log", ".", "debug", "(", "\"HAS LANE RELS: %s\"", "%", "self", ".", "current", ".", "lane_relations", ")", "try", ":", "cond_result", "=", "eval", "(", "self", ".", "current", ".", "lane_relations", ",", "context", ")", "except", ":", "log", ".", "exception", "(", "\"CONDITION EVAL ERROR : %s || %s\"", "%", "(", "self", ".", "current", ".", "lane_relations", ",", "context", ")", ")", "raise", "if", "not", "cond_result", ":", "log", ".", "debug", "(", "\"LANE RELATION ERR: %s %s\"", "%", "(", "self", ".", "current", ".", "lane_relations", ",", "context", ")", ")", "raise", "HTTPError", "(", "403", ",", "\"You aren't qualified for this lane: %s\"", "%", "self", ".", "current", ".", "lane_relations", ")" ]
One or more permissions can be associated with a lane of a workflow. In a similar way, a lane can be restricted with relation to other lanes of the workflow. This method called on lane changes and checks user has required permissions and relations. Raises: HTTPForbidden: if the current user hasn't got the required permissions and proper relations
[ "One", "or", "more", "permissions", "can", "be", "associated", "with", "a", "lane", "of", "a", "workflow", ".", "In", "a", "similar", "way", "a", "lane", "can", "be", "restricted", "with", "relation", "to", "other", "lanes", "of", "the", "workflow", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L657-L690
zetaops/zengine
zengine/engine.py
ZEngine.check_for_permission
def check_for_permission(self): # TODO: Works but not beautiful, needs review! """ Checks if current user (or role) has the required permission for current workflow step. Raises: HTTPError: if user doesn't have required permissions. """ if self.current.task: lane = self.current.lane_id permission = "%s.%s.%s" % (self.current.workflow_name, lane, self.current.task_name) else: permission = self.current.workflow_name log.debug("CHECK PERM: %s" % permission) if (self.current.task_type not in PERM_REQ_TASK_TYPES or permission.startswith(tuple(settings.ANONYMOUS_WORKFLOWS)) or (self.current.is_auth and permission.startswith(tuple(settings.COMMON_WORKFLOWS)))): return # FIXME:needs hardening log.debug("REQUIRE PERM: %s" % permission) if not self.current.has_permission(permission): raise HTTPError(403, "You don't have required permission: %s" % permission)
python
def check_for_permission(self): # TODO: Works but not beautiful, needs review! """ Checks if current user (or role) has the required permission for current workflow step. Raises: HTTPError: if user doesn't have required permissions. """ if self.current.task: lane = self.current.lane_id permission = "%s.%s.%s" % (self.current.workflow_name, lane, self.current.task_name) else: permission = self.current.workflow_name log.debug("CHECK PERM: %s" % permission) if (self.current.task_type not in PERM_REQ_TASK_TYPES or permission.startswith(tuple(settings.ANONYMOUS_WORKFLOWS)) or (self.current.is_auth and permission.startswith(tuple(settings.COMMON_WORKFLOWS)))): return # FIXME:needs hardening log.debug("REQUIRE PERM: %s" % permission) if not self.current.has_permission(permission): raise HTTPError(403, "You don't have required permission: %s" % permission)
[ "def", "check_for_permission", "(", "self", ")", ":", "# TODO: Works but not beautiful, needs review!", "if", "self", ".", "current", ".", "task", ":", "lane", "=", "self", ".", "current", ".", "lane_id", "permission", "=", "\"%s.%s.%s\"", "%", "(", "self", ".", "current", ".", "workflow_name", ",", "lane", ",", "self", ".", "current", ".", "task_name", ")", "else", ":", "permission", "=", "self", ".", "current", ".", "workflow_name", "log", ".", "debug", "(", "\"CHECK PERM: %s\"", "%", "permission", ")", "if", "(", "self", ".", "current", ".", "task_type", "not", "in", "PERM_REQ_TASK_TYPES", "or", "permission", ".", "startswith", "(", "tuple", "(", "settings", ".", "ANONYMOUS_WORKFLOWS", ")", ")", "or", "(", "self", ".", "current", ".", "is_auth", "and", "permission", ".", "startswith", "(", "tuple", "(", "settings", ".", "COMMON_WORKFLOWS", ")", ")", ")", ")", ":", "return", "# FIXME:needs hardening", "log", ".", "debug", "(", "\"REQUIRE PERM: %s\"", "%", "permission", ")", "if", "not", "self", ".", "current", ".", "has_permission", "(", "permission", ")", ":", "raise", "HTTPError", "(", "403", ",", "\"You don't have required permission: %s\"", "%", "permission", ")" ]
Checks if current user (or role) has the required permission for current workflow step. Raises: HTTPError: if user doesn't have required permissions.
[ "Checks", "if", "current", "user", "(", "or", "role", ")", "has", "the", "required", "permission", "for", "current", "workflow", "step", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L692-L716
zetaops/zengine
zengine/engine.py
ZEngine.handle_wf_finalization
def handle_wf_finalization(self): """ Removes the ``token`` key from ``current.output`` if WF is over. """ if ((not self.current.flow_enabled or ( self.current.task_type.startswith('End') and not self.are_we_in_subprocess())) and 'token' in self.current.output): del self.current.output['token']
python
def handle_wf_finalization(self): """ Removes the ``token`` key from ``current.output`` if WF is over. """ if ((not self.current.flow_enabled or ( self.current.task_type.startswith('End') and not self.are_we_in_subprocess())) and 'token' in self.current.output): del self.current.output['token']
[ "def", "handle_wf_finalization", "(", "self", ")", ":", "if", "(", "(", "not", "self", ".", "current", ".", "flow_enabled", "or", "(", "self", ".", "current", ".", "task_type", ".", "startswith", "(", "'End'", ")", "and", "not", "self", ".", "are_we_in_subprocess", "(", ")", ")", ")", "and", "'token'", "in", "self", ".", "current", ".", "output", ")", ":", "del", "self", ".", "current", ".", "output", "[", "'token'", "]" ]
Removes the ``token`` key from ``current.output`` if WF is over.
[ "Removes", "the", "token", "key", "from", "current", ".", "output", "if", "WF", "is", "over", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L718-L725
cimm-kzn/CGRtools
CGRtools/utils/rdkit.py
from_rdkit_molecule
def from_rdkit_molecule(data): """ RDKit molecule object to MoleculeContainer converter """ m = MoleculeContainer() atoms, mapping = [], [] for a in data.GetAtoms(): atom = {'element': a.GetSymbol(), 'charge': a.GetFormalCharge()} atoms.append(atom) mapping.append(a.GetAtomMapNum()) isotope = a.GetIsotope() if isotope: atom['isotope'] = isotope radical = a.GetNumRadicalElectrons() if radical: atom['multiplicity'] = radical + 1 conformers = data.GetConformers() if conformers: for atom, (x, y, z) in zip(atoms, conformers[0].GetPositions()): atom['x'] = x atom['y'] = y atom['z'] = z for atom, mapping in zip(atoms, mapping): a = m.add_atom(atom) if mapping: m.atom(a)._parsed_mapping = mapping for bond in data.GetBonds(): m.add_bond(bond.GetBeginAtomIdx() + 1, bond.GetEndAtomIdx() + 1, _rdkit_bond_map[bond.GetBondType()]) return m
python
def from_rdkit_molecule(data): """ RDKit molecule object to MoleculeContainer converter """ m = MoleculeContainer() atoms, mapping = [], [] for a in data.GetAtoms(): atom = {'element': a.GetSymbol(), 'charge': a.GetFormalCharge()} atoms.append(atom) mapping.append(a.GetAtomMapNum()) isotope = a.GetIsotope() if isotope: atom['isotope'] = isotope radical = a.GetNumRadicalElectrons() if radical: atom['multiplicity'] = radical + 1 conformers = data.GetConformers() if conformers: for atom, (x, y, z) in zip(atoms, conformers[0].GetPositions()): atom['x'] = x atom['y'] = y atom['z'] = z for atom, mapping in zip(atoms, mapping): a = m.add_atom(atom) if mapping: m.atom(a)._parsed_mapping = mapping for bond in data.GetBonds(): m.add_bond(bond.GetBeginAtomIdx() + 1, bond.GetEndAtomIdx() + 1, _rdkit_bond_map[bond.GetBondType()]) return m
[ "def", "from_rdkit_molecule", "(", "data", ")", ":", "m", "=", "MoleculeContainer", "(", ")", "atoms", ",", "mapping", "=", "[", "]", ",", "[", "]", "for", "a", "in", "data", ".", "GetAtoms", "(", ")", ":", "atom", "=", "{", "'element'", ":", "a", ".", "GetSymbol", "(", ")", ",", "'charge'", ":", "a", ".", "GetFormalCharge", "(", ")", "}", "atoms", ".", "append", "(", "atom", ")", "mapping", ".", "append", "(", "a", ".", "GetAtomMapNum", "(", ")", ")", "isotope", "=", "a", ".", "GetIsotope", "(", ")", "if", "isotope", ":", "atom", "[", "'isotope'", "]", "=", "isotope", "radical", "=", "a", ".", "GetNumRadicalElectrons", "(", ")", "if", "radical", ":", "atom", "[", "'multiplicity'", "]", "=", "radical", "+", "1", "conformers", "=", "data", ".", "GetConformers", "(", ")", "if", "conformers", ":", "for", "atom", ",", "(", "x", ",", "y", ",", "z", ")", "in", "zip", "(", "atoms", ",", "conformers", "[", "0", "]", ".", "GetPositions", "(", ")", ")", ":", "atom", "[", "'x'", "]", "=", "x", "atom", "[", "'y'", "]", "=", "y", "atom", "[", "'z'", "]", "=", "z", "for", "atom", ",", "mapping", "in", "zip", "(", "atoms", ",", "mapping", ")", ":", "a", "=", "m", ".", "add_atom", "(", "atom", ")", "if", "mapping", ":", "m", ".", "atom", "(", "a", ")", ".", "_parsed_mapping", "=", "mapping", "for", "bond", "in", "data", ".", "GetBonds", "(", ")", ":", "m", ".", "add_bond", "(", "bond", ".", "GetBeginAtomIdx", "(", ")", "+", "1", ",", "bond", ".", "GetEndAtomIdx", "(", ")", "+", "1", ",", "_rdkit_bond_map", "[", "bond", ".", "GetBondType", "(", ")", "]", ")", "return", "m" ]
RDKit molecule object to MoleculeContainer converter
[ "RDKit", "molecule", "object", "to", "MoleculeContainer", "converter" ]
train
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/utils/rdkit.py#L23-L56
cimm-kzn/CGRtools
CGRtools/utils/rdkit.py
to_rdkit_molecule
def to_rdkit_molecule(data): """ MoleculeContainer to RDKit molecule object converter """ mol = RWMol() conf = Conformer() mapping = {} is_3d = False for n, a in data.atoms(): ra = Atom(a.number) ra.SetAtomMapNum(n) if a.charge: ra.SetFormalCharge(a.charge) if a.isotope != a.common_isotope: ra.SetIsotope(a.isotope) if a.radical: ra.SetNumRadicalElectrons(a.radical) mapping[n] = m = mol.AddAtom(ra) conf.SetAtomPosition(m, (a.x, a.y, a.z)) if a.z: is_3d = True if not is_3d: conf.Set3D(False) for n, m, b in data.bonds(): mol.AddBond(mapping[n], mapping[m], _bond_map[b.order]) mol.AddConformer(conf) SanitizeMol(mol) return mol
python
def to_rdkit_molecule(data): """ MoleculeContainer to RDKit molecule object converter """ mol = RWMol() conf = Conformer() mapping = {} is_3d = False for n, a in data.atoms(): ra = Atom(a.number) ra.SetAtomMapNum(n) if a.charge: ra.SetFormalCharge(a.charge) if a.isotope != a.common_isotope: ra.SetIsotope(a.isotope) if a.radical: ra.SetNumRadicalElectrons(a.radical) mapping[n] = m = mol.AddAtom(ra) conf.SetAtomPosition(m, (a.x, a.y, a.z)) if a.z: is_3d = True if not is_3d: conf.Set3D(False) for n, m, b in data.bonds(): mol.AddBond(mapping[n], mapping[m], _bond_map[b.order]) mol.AddConformer(conf) SanitizeMol(mol) return mol
[ "def", "to_rdkit_molecule", "(", "data", ")", ":", "mol", "=", "RWMol", "(", ")", "conf", "=", "Conformer", "(", ")", "mapping", "=", "{", "}", "is_3d", "=", "False", "for", "n", ",", "a", "in", "data", ".", "atoms", "(", ")", ":", "ra", "=", "Atom", "(", "a", ".", "number", ")", "ra", ".", "SetAtomMapNum", "(", "n", ")", "if", "a", ".", "charge", ":", "ra", ".", "SetFormalCharge", "(", "a", ".", "charge", ")", "if", "a", ".", "isotope", "!=", "a", ".", "common_isotope", ":", "ra", ".", "SetIsotope", "(", "a", ".", "isotope", ")", "if", "a", ".", "radical", ":", "ra", ".", "SetNumRadicalElectrons", "(", "a", ".", "radical", ")", "mapping", "[", "n", "]", "=", "m", "=", "mol", ".", "AddAtom", "(", "ra", ")", "conf", ".", "SetAtomPosition", "(", "m", ",", "(", "a", ".", "x", ",", "a", ".", "y", ",", "a", ".", "z", ")", ")", "if", "a", ".", "z", ":", "is_3d", "=", "True", "if", "not", "is_3d", ":", "conf", ".", "Set3D", "(", "False", ")", "for", "n", ",", "m", ",", "b", "in", "data", ".", "bonds", "(", ")", ":", "mol", ".", "AddBond", "(", "mapping", "[", "n", "]", ",", "mapping", "[", "m", "]", ",", "_bond_map", "[", "b", ".", "order", "]", ")", "mol", ".", "AddConformer", "(", "conf", ")", "SanitizeMol", "(", "mol", ")", "return", "mol" ]
MoleculeContainer to RDKit molecule object converter
[ "MoleculeContainer", "to", "RDKit", "molecule", "object", "converter" ]
train
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/utils/rdkit.py#L59-L88
cimm-kzn/CGRtools
CGRtools/algorithms/strings.py
StringCommon.__dfs
def __dfs(self, start, weights, depth_limit): """ modified NX dfs """ adj = self._adj stack = [(start, depth_limit, iter(sorted(adj[start], key=weights)))] visited = {start} disconnected = defaultdict(list) edges = defaultdict(list) while stack: parent, depth_now, children = stack[-1] try: child = next(children) except StopIteration: stack.pop() else: if child not in visited: edges[parent].append(child) visited.add(child) if depth_now > 1: front = adj[child].keys() - {parent} if front: stack.append((child, depth_now - 1, iter(sorted(front, key=weights)))) elif child not in disconnected: disconnected[parent].append(child) return visited, edges, disconnected
python
def __dfs(self, start, weights, depth_limit): """ modified NX dfs """ adj = self._adj stack = [(start, depth_limit, iter(sorted(adj[start], key=weights)))] visited = {start} disconnected = defaultdict(list) edges = defaultdict(list) while stack: parent, depth_now, children = stack[-1] try: child = next(children) except StopIteration: stack.pop() else: if child not in visited: edges[parent].append(child) visited.add(child) if depth_now > 1: front = adj[child].keys() - {parent} if front: stack.append((child, depth_now - 1, iter(sorted(front, key=weights)))) elif child not in disconnected: disconnected[parent].append(child) return visited, edges, disconnected
[ "def", "__dfs", "(", "self", ",", "start", ",", "weights", ",", "depth_limit", ")", ":", "adj", "=", "self", ".", "_adj", "stack", "=", "[", "(", "start", ",", "depth_limit", ",", "iter", "(", "sorted", "(", "adj", "[", "start", "]", ",", "key", "=", "weights", ")", ")", ")", "]", "visited", "=", "{", "start", "}", "disconnected", "=", "defaultdict", "(", "list", ")", "edges", "=", "defaultdict", "(", "list", ")", "while", "stack", ":", "parent", ",", "depth_now", ",", "children", "=", "stack", "[", "-", "1", "]", "try", ":", "child", "=", "next", "(", "children", ")", "except", "StopIteration", ":", "stack", ".", "pop", "(", ")", "else", ":", "if", "child", "not", "in", "visited", ":", "edges", "[", "parent", "]", ".", "append", "(", "child", ")", "visited", ".", "add", "(", "child", ")", "if", "depth_now", ">", "1", ":", "front", "=", "adj", "[", "child", "]", ".", "keys", "(", ")", "-", "{", "parent", "}", "if", "front", ":", "stack", ".", "append", "(", "(", "child", ",", "depth_now", "-", "1", ",", "iter", "(", "sorted", "(", "front", ",", "key", "=", "weights", ")", ")", ")", ")", "elif", "child", "not", "in", "disconnected", ":", "disconnected", "[", "parent", "]", ".", "append", "(", "child", ")", "return", "visited", ",", "edges", ",", "disconnected" ]
modified NX dfs
[ "modified", "NX", "dfs" ]
train
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/algorithms/strings.py#L130-L158
camptocamp/marabunta
marabunta/config.py
get_args_parser
def get_args_parser(): """Return a parser for command line options.""" parser = argparse.ArgumentParser( description='Marabunta: Migrating ants for Odoo') parser.add_argument('--migration-file', '-f', action=EnvDefault, envvar='MARABUNTA_MIGRATION_FILE', required=True, help='The yaml file containing the migration steps') parser.add_argument('--database', '-d', action=EnvDefault, envvar='MARABUNTA_DATABASE', required=True, help="Odoo's database") parser.add_argument('--db-user', '-u', action=EnvDefault, envvar='MARABUNTA_DB_USER', required=True, help="Odoo's database user") parser.add_argument('--db-password', '-w', action=EnvDefault, envvar='MARABUNTA_DB_PASSWORD', required=True, help="Odoo's database password") parser.add_argument('--db-port', '-p', default=os.environ.get('MARABUNTA_DB_PORT', 5432), help="Odoo's database port") parser.add_argument('--db-host', '-H', default=os.environ.get('MARABUNTA_DB_HOST', 'localhost'), help="Odoo's database host") parser.add_argument('--mode', action=EnvDefault, envvar='MARABUNTA_MODE', required=False, help="Specify the mode in which we run the migration," "such as 'demo' or 'prod'. Additional operations " "of this mode will be executed after the main " "operations and the addons list of this mode " "will be merged with the main addons list.") parser.add_argument('--allow-serie', action=BoolEnvDefault, required=False, envvar='MARABUNTA_ALLOW_SERIE', help='Allow to run more than 1 version upgrade at a ' 'time.') parser.add_argument('--force-version', required=False, default=os.environ.get('MARABUNTA_FORCE_VERSION'), help='Force upgrade of a version, even if it has ' 'already been applied.') group = parser.add_argument_group( title='Web', description='Configuration related to the internal web server, ' 'used to publish a maintenance page during the migration.', ) group.add_argument('--web-host', required=False, default=os.environ.get('MARABUNTA_WEB_HOST', '0.0.0.0'), help='Host for the web server') group.add_argument('--web-port', required=False, default=os.environ.get('MARABUNTA_WEB_PORT', 8069), help='Port for the web server') group.add_argument('--web-custom-html', required=False, default=os.environ.get( 'MARABUNTA_WEB_CUSTOM_HTML' ), help='Path to a custom html file to publish') return parser
python
def get_args_parser(): """Return a parser for command line options.""" parser = argparse.ArgumentParser( description='Marabunta: Migrating ants for Odoo') parser.add_argument('--migration-file', '-f', action=EnvDefault, envvar='MARABUNTA_MIGRATION_FILE', required=True, help='The yaml file containing the migration steps') parser.add_argument('--database', '-d', action=EnvDefault, envvar='MARABUNTA_DATABASE', required=True, help="Odoo's database") parser.add_argument('--db-user', '-u', action=EnvDefault, envvar='MARABUNTA_DB_USER', required=True, help="Odoo's database user") parser.add_argument('--db-password', '-w', action=EnvDefault, envvar='MARABUNTA_DB_PASSWORD', required=True, help="Odoo's database password") parser.add_argument('--db-port', '-p', default=os.environ.get('MARABUNTA_DB_PORT', 5432), help="Odoo's database port") parser.add_argument('--db-host', '-H', default=os.environ.get('MARABUNTA_DB_HOST', 'localhost'), help="Odoo's database host") parser.add_argument('--mode', action=EnvDefault, envvar='MARABUNTA_MODE', required=False, help="Specify the mode in which we run the migration," "such as 'demo' or 'prod'. Additional operations " "of this mode will be executed after the main " "operations and the addons list of this mode " "will be merged with the main addons list.") parser.add_argument('--allow-serie', action=BoolEnvDefault, required=False, envvar='MARABUNTA_ALLOW_SERIE', help='Allow to run more than 1 version upgrade at a ' 'time.') parser.add_argument('--force-version', required=False, default=os.environ.get('MARABUNTA_FORCE_VERSION'), help='Force upgrade of a version, even if it has ' 'already been applied.') group = parser.add_argument_group( title='Web', description='Configuration related to the internal web server, ' 'used to publish a maintenance page during the migration.', ) group.add_argument('--web-host', required=False, default=os.environ.get('MARABUNTA_WEB_HOST', '0.0.0.0'), help='Host for the web server') group.add_argument('--web-port', required=False, default=os.environ.get('MARABUNTA_WEB_PORT', 8069), help='Port for the web server') group.add_argument('--web-custom-html', required=False, default=os.environ.get( 'MARABUNTA_WEB_CUSTOM_HTML' ), help='Path to a custom html file to publish') return parser
[ "def", "get_args_parser", "(", ")", ":", "parser", "=", "argparse", ".", "ArgumentParser", "(", "description", "=", "'Marabunta: Migrating ants for Odoo'", ")", "parser", ".", "add_argument", "(", "'--migration-file'", ",", "'-f'", ",", "action", "=", "EnvDefault", ",", "envvar", "=", "'MARABUNTA_MIGRATION_FILE'", ",", "required", "=", "True", ",", "help", "=", "'The yaml file containing the migration steps'", ")", "parser", ".", "add_argument", "(", "'--database'", ",", "'-d'", ",", "action", "=", "EnvDefault", ",", "envvar", "=", "'MARABUNTA_DATABASE'", ",", "required", "=", "True", ",", "help", "=", "\"Odoo's database\"", ")", "parser", ".", "add_argument", "(", "'--db-user'", ",", "'-u'", ",", "action", "=", "EnvDefault", ",", "envvar", "=", "'MARABUNTA_DB_USER'", ",", "required", "=", "True", ",", "help", "=", "\"Odoo's database user\"", ")", "parser", ".", "add_argument", "(", "'--db-password'", ",", "'-w'", ",", "action", "=", "EnvDefault", ",", "envvar", "=", "'MARABUNTA_DB_PASSWORD'", ",", "required", "=", "True", ",", "help", "=", "\"Odoo's database password\"", ")", "parser", ".", "add_argument", "(", "'--db-port'", ",", "'-p'", ",", "default", "=", "os", ".", "environ", ".", "get", "(", "'MARABUNTA_DB_PORT'", ",", "5432", ")", ",", "help", "=", "\"Odoo's database port\"", ")", "parser", ".", "add_argument", "(", "'--db-host'", ",", "'-H'", ",", "default", "=", "os", ".", "environ", ".", "get", "(", "'MARABUNTA_DB_HOST'", ",", "'localhost'", ")", ",", "help", "=", "\"Odoo's database host\"", ")", "parser", ".", "add_argument", "(", "'--mode'", ",", "action", "=", "EnvDefault", ",", "envvar", "=", "'MARABUNTA_MODE'", ",", "required", "=", "False", ",", "help", "=", "\"Specify the mode in which we run the migration,\"", "\"such as 'demo' or 'prod'. Additional operations \"", "\"of this mode will be executed after the main \"", "\"operations and the addons list of this mode \"", "\"will be merged with the main addons list.\"", ")", "parser", ".", "add_argument", "(", "'--allow-serie'", ",", "action", "=", "BoolEnvDefault", ",", "required", "=", "False", ",", "envvar", "=", "'MARABUNTA_ALLOW_SERIE'", ",", "help", "=", "'Allow to run more than 1 version upgrade at a '", "'time.'", ")", "parser", ".", "add_argument", "(", "'--force-version'", ",", "required", "=", "False", ",", "default", "=", "os", ".", "environ", ".", "get", "(", "'MARABUNTA_FORCE_VERSION'", ")", ",", "help", "=", "'Force upgrade of a version, even if it has '", "'already been applied.'", ")", "group", "=", "parser", ".", "add_argument_group", "(", "title", "=", "'Web'", ",", "description", "=", "'Configuration related to the internal web server, '", "'used to publish a maintenance page during the migration.'", ",", ")", "group", ".", "add_argument", "(", "'--web-host'", ",", "required", "=", "False", ",", "default", "=", "os", ".", "environ", ".", "get", "(", "'MARABUNTA_WEB_HOST'", ",", "'0.0.0.0'", ")", ",", "help", "=", "'Host for the web server'", ")", "group", ".", "add_argument", "(", "'--web-port'", ",", "required", "=", "False", ",", "default", "=", "os", ".", "environ", ".", "get", "(", "'MARABUNTA_WEB_PORT'", ",", "8069", ")", ",", "help", "=", "'Port for the web server'", ")", "group", ".", "add_argument", "(", "'--web-custom-html'", ",", "required", "=", "False", ",", "default", "=", "os", ".", "environ", ".", "get", "(", "'MARABUNTA_WEB_CUSTOM_HTML'", ")", ",", "help", "=", "'Path to a custom html file to publish'", ")", "return", "parser" ]
Return a parser for command line options.
[ "Return", "a", "parser", "for", "command", "line", "options", "." ]
train
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/config.py#L90-L161
camptocamp/marabunta
marabunta/config.py
Config.from_parse_args
def from_parse_args(cls, args): """Constructor from command line args. :param args: parse command line arguments :type args: argparse.ArgumentParser """ return cls(args.migration_file, args.database, db_user=args.db_user, db_password=args.db_password, db_port=args.db_port, db_host=args.db_host, mode=args.mode, allow_serie=args.allow_serie, force_version=args.force_version, web_host=args.web_host, web_port=args.web_port, web_custom_html=args.web_custom_html, )
python
def from_parse_args(cls, args): """Constructor from command line args. :param args: parse command line arguments :type args: argparse.ArgumentParser """ return cls(args.migration_file, args.database, db_user=args.db_user, db_password=args.db_password, db_port=args.db_port, db_host=args.db_host, mode=args.mode, allow_serie=args.allow_serie, force_version=args.force_version, web_host=args.web_host, web_port=args.web_port, web_custom_html=args.web_custom_html, )
[ "def", "from_parse_args", "(", "cls", ",", "args", ")", ":", "return", "cls", "(", "args", ".", "migration_file", ",", "args", ".", "database", ",", "db_user", "=", "args", ".", "db_user", ",", "db_password", "=", "args", ".", "db_password", ",", "db_port", "=", "args", ".", "db_port", ",", "db_host", "=", "args", ".", "db_host", ",", "mode", "=", "args", ".", "mode", ",", "allow_serie", "=", "args", ".", "allow_serie", ",", "force_version", "=", "args", ".", "force_version", ",", "web_host", "=", "args", ".", "web_host", ",", "web_port", "=", "args", ".", "web_port", ",", "web_custom_html", "=", "args", ".", "web_custom_html", ",", ")" ]
Constructor from command line args. :param args: parse command line arguments :type args: argparse.ArgumentParser
[ "Constructor", "from", "command", "line", "args", "." ]
train
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/config.py#L40-L60
zetaops/zengine
zengine/views/base.py
BaseView.set_current
def set_current(self, current): """ Creates some aliases for attributes of ``current``. Args: current: :attr:`~zengine.engine.WFCurrent` object. """ self.current = current self.input = current.input # self.req = current.request # self.resp = current.response self.output = current.output self.cmd = current.task_data['cmd'] if self.cmd and NEXT_CMD_SPLITTER in self.cmd: self.cmd, self.next_cmd = self.cmd.split(NEXT_CMD_SPLITTER) else: self.next_cmd = None
python
def set_current(self, current): """ Creates some aliases for attributes of ``current``. Args: current: :attr:`~zengine.engine.WFCurrent` object. """ self.current = current self.input = current.input # self.req = current.request # self.resp = current.response self.output = current.output self.cmd = current.task_data['cmd'] if self.cmd and NEXT_CMD_SPLITTER in self.cmd: self.cmd, self.next_cmd = self.cmd.split(NEXT_CMD_SPLITTER) else: self.next_cmd = None
[ "def", "set_current", "(", "self", ",", "current", ")", ":", "self", ".", "current", "=", "current", "self", ".", "input", "=", "current", ".", "input", "# self.req = current.request", "# self.resp = current.response", "self", ".", "output", "=", "current", ".", "output", "self", ".", "cmd", "=", "current", ".", "task_data", "[", "'cmd'", "]", "if", "self", ".", "cmd", "and", "NEXT_CMD_SPLITTER", "in", "self", ".", "cmd", ":", "self", ".", "cmd", ",", "self", ".", "next_cmd", "=", "self", ".", "cmd", ".", "split", "(", "NEXT_CMD_SPLITTER", ")", "else", ":", "self", ".", "next_cmd", "=", "None" ]
Creates some aliases for attributes of ``current``. Args: current: :attr:`~zengine.engine.WFCurrent` object.
[ "Creates", "some", "aliases", "for", "attributes", "of", "current", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/base.py#L35-L52
zetaops/zengine
zengine/views/base.py
BaseView.form_out
def form_out(self, _form=None): """ Renders form. Applies form modifiers, then writes result to response payload. If supplied, given form object instance will be used instead of view's default ObjectForm. Args: _form (:py:attr:`~zengine.forms.json_form.JsonForm`): Form object to override `self.object_form` """ _form = _form or self.object_form self.output['forms'] = _form.serialize() self._add_meta_props(_form) self.output['forms']['grouping'] = _form.Meta.grouping self.output['forms']['constraints'] = _form.Meta.constraints self._patch_form(self.output['forms']) self.set_client_cmd('form')
python
def form_out(self, _form=None): """ Renders form. Applies form modifiers, then writes result to response payload. If supplied, given form object instance will be used instead of view's default ObjectForm. Args: _form (:py:attr:`~zengine.forms.json_form.JsonForm`): Form object to override `self.object_form` """ _form = _form or self.object_form self.output['forms'] = _form.serialize() self._add_meta_props(_form) self.output['forms']['grouping'] = _form.Meta.grouping self.output['forms']['constraints'] = _form.Meta.constraints self._patch_form(self.output['forms']) self.set_client_cmd('form')
[ "def", "form_out", "(", "self", ",", "_form", "=", "None", ")", ":", "_form", "=", "_form", "or", "self", ".", "object_form", "self", ".", "output", "[", "'forms'", "]", "=", "_form", ".", "serialize", "(", ")", "self", ".", "_add_meta_props", "(", "_form", ")", "self", ".", "output", "[", "'forms'", "]", "[", "'grouping'", "]", "=", "_form", ".", "Meta", ".", "grouping", "self", ".", "output", "[", "'forms'", "]", "[", "'constraints'", "]", "=", "_form", ".", "Meta", ".", "constraints", "self", ".", "_patch_form", "(", "self", ".", "output", "[", "'forms'", "]", ")", "self", ".", "set_client_cmd", "(", "'form'", ")" ]
Renders form. Applies form modifiers, then writes result to response payload. If supplied, given form object instance will be used instead of view's default ObjectForm. Args: _form (:py:attr:`~zengine.forms.json_form.JsonForm`): Form object to override `self.object_form`
[ "Renders", "form", ".", "Applies", "form", "modifiers", "then", "writes", "result", "to", "response", "payload", ".", "If", "supplied", "given", "form", "object", "instance", "will", "be", "used", "instead", "of", "view", "s", "default", "ObjectForm", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/base.py#L86-L103
zetaops/zengine
zengine/views/base.py
BaseView.set_client_cmd
def set_client_cmd(self, *args): """ Adds given cmd(s) to ``self.output['client_cmd']`` Args: *args: Client commands. """ self.client_cmd.update(args) self.output['client_cmd'] = list(self.client_cmd)
python
def set_client_cmd(self, *args): """ Adds given cmd(s) to ``self.output['client_cmd']`` Args: *args: Client commands. """ self.client_cmd.update(args) self.output['client_cmd'] = list(self.client_cmd)
[ "def", "set_client_cmd", "(", "self", ",", "*", "args", ")", ":", "self", ".", "client_cmd", ".", "update", "(", "args", ")", "self", ".", "output", "[", "'client_cmd'", "]", "=", "list", "(", "self", ".", "client_cmd", ")" ]
Adds given cmd(s) to ``self.output['client_cmd']`` Args: *args: Client commands.
[ "Adds", "given", "cmd", "(", "s", ")", "to", "self", ".", "output", "[", "client_cmd", "]" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/base.py#L117-L125
zetaops/zengine
zengine/management_commands.py
UpdatePermissions.run
def run(self): """ Creates new permissions. """ from pyoko.lib.utils import get_object_from_path from zengine.config import settings model = get_object_from_path(settings.PERMISSION_MODEL) perm_provider = get_object_from_path(settings.PERMISSION_PROVIDER) existing_perms = [] new_perms = [] for code, name, desc in perm_provider(): code = six.text_type(code) if self.manager.args.dry: exists = model.objects.filter(code=code, name=name) if exists: perm = exists[0] new = False else: new = True perm = model(code=code, name=name) else: try: perm = model.objects.get(code) existing_perms.append(perm) except ObjectDoesNotExist: perm = model(description=desc, code=code, name=name) perm.key = code perm.save() new_perms.append(perm) # perm, new = model.objects.get_or_create({'description': desc}, code=code, name=name) # if new: # new_perms.append(perm) # else: # existing_perms.append(perm) report = "\n\n%s permission(s) were found in DB. " % len(existing_perms) if new_perms: report += "\n%s new permission record added. " % len(new_perms) else: report += 'No new perms added. ' if new_perms: if not self.manager.args.dry: SelectBoxCache.flush(model.__name__) report += 'Total %s perms exists.' % (len(existing_perms) + len(new_perms)) report = "\n + " + "\n + ".join([p.name or p.code for p in new_perms]) + report if self.manager.args.dry: print("\n~~~~~~~~~~~~~~ DRY RUN ~~~~~~~~~~~~~~\n") print(report + "\n")
python
def run(self): """ Creates new permissions. """ from pyoko.lib.utils import get_object_from_path from zengine.config import settings model = get_object_from_path(settings.PERMISSION_MODEL) perm_provider = get_object_from_path(settings.PERMISSION_PROVIDER) existing_perms = [] new_perms = [] for code, name, desc in perm_provider(): code = six.text_type(code) if self.manager.args.dry: exists = model.objects.filter(code=code, name=name) if exists: perm = exists[0] new = False else: new = True perm = model(code=code, name=name) else: try: perm = model.objects.get(code) existing_perms.append(perm) except ObjectDoesNotExist: perm = model(description=desc, code=code, name=name) perm.key = code perm.save() new_perms.append(perm) # perm, new = model.objects.get_or_create({'description': desc}, code=code, name=name) # if new: # new_perms.append(perm) # else: # existing_perms.append(perm) report = "\n\n%s permission(s) were found in DB. " % len(existing_perms) if new_perms: report += "\n%s new permission record added. " % len(new_perms) else: report += 'No new perms added. ' if new_perms: if not self.manager.args.dry: SelectBoxCache.flush(model.__name__) report += 'Total %s perms exists.' % (len(existing_perms) + len(new_perms)) report = "\n + " + "\n + ".join([p.name or p.code for p in new_perms]) + report if self.manager.args.dry: print("\n~~~~~~~~~~~~~~ DRY RUN ~~~~~~~~~~~~~~\n") print(report + "\n")
[ "def", "run", "(", "self", ")", ":", "from", "pyoko", ".", "lib", ".", "utils", "import", "get_object_from_path", "from", "zengine", ".", "config", "import", "settings", "model", "=", "get_object_from_path", "(", "settings", ".", "PERMISSION_MODEL", ")", "perm_provider", "=", "get_object_from_path", "(", "settings", ".", "PERMISSION_PROVIDER", ")", "existing_perms", "=", "[", "]", "new_perms", "=", "[", "]", "for", "code", ",", "name", ",", "desc", "in", "perm_provider", "(", ")", ":", "code", "=", "six", ".", "text_type", "(", "code", ")", "if", "self", ".", "manager", ".", "args", ".", "dry", ":", "exists", "=", "model", ".", "objects", ".", "filter", "(", "code", "=", "code", ",", "name", "=", "name", ")", "if", "exists", ":", "perm", "=", "exists", "[", "0", "]", "new", "=", "False", "else", ":", "new", "=", "True", "perm", "=", "model", "(", "code", "=", "code", ",", "name", "=", "name", ")", "else", ":", "try", ":", "perm", "=", "model", ".", "objects", ".", "get", "(", "code", ")", "existing_perms", ".", "append", "(", "perm", ")", "except", "ObjectDoesNotExist", ":", "perm", "=", "model", "(", "description", "=", "desc", ",", "code", "=", "code", ",", "name", "=", "name", ")", "perm", ".", "key", "=", "code", "perm", ".", "save", "(", ")", "new_perms", ".", "append", "(", "perm", ")", "# perm, new = model.objects.get_or_create({'description': desc}, code=code, name=name)", "# if new:", "# new_perms.append(perm)", "# else:", "# existing_perms.append(perm)", "report", "=", "\"\\n\\n%s permission(s) were found in DB. \"", "%", "len", "(", "existing_perms", ")", "if", "new_perms", ":", "report", "+=", "\"\\n%s new permission record added. \"", "%", "len", "(", "new_perms", ")", "else", ":", "report", "+=", "'No new perms added. '", "if", "new_perms", ":", "if", "not", "self", ".", "manager", ".", "args", ".", "dry", ":", "SelectBoxCache", ".", "flush", "(", "model", ".", "__name__", ")", "report", "+=", "'Total %s perms exists.'", "%", "(", "len", "(", "existing_perms", ")", "+", "len", "(", "new_perms", ")", ")", "report", "=", "\"\\n + \"", "+", "\"\\n + \"", ".", "join", "(", "[", "p", ".", "name", "or", "p", ".", "code", "for", "p", "in", "new_perms", "]", ")", "+", "report", "if", "self", ".", "manager", ".", "args", ".", "dry", ":", "print", "(", "\"\\n~~~~~~~~~~~~~~ DRY RUN ~~~~~~~~~~~~~~\\n\"", ")", "print", "(", "report", "+", "\"\\n\"", ")" ]
Creates new permissions.
[ "Creates", "new", "permissions", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L44-L92
zetaops/zengine
zengine/management_commands.py
CreateUser.run
def run(self): """ Creates user, encrypts password. """ from zengine.models import User user = User(username=self.manager.args.username, superuser=self.manager.args.super) user.set_password(self.manager.args.password) user.save() print("New user created with ID: %s" % user.key)
python
def run(self): """ Creates user, encrypts password. """ from zengine.models import User user = User(username=self.manager.args.username, superuser=self.manager.args.super) user.set_password(self.manager.args.password) user.save() print("New user created with ID: %s" % user.key)
[ "def", "run", "(", "self", ")", ":", "from", "zengine", ".", "models", "import", "User", "user", "=", "User", "(", "username", "=", "self", ".", "manager", ".", "args", ".", "username", ",", "superuser", "=", "self", ".", "manager", ".", "args", ".", "super", ")", "user", ".", "set_password", "(", "self", ".", "manager", ".", "args", ".", "password", ")", "user", ".", "save", "(", ")", "print", "(", "\"New user created with ID: %s\"", "%", "user", ".", "key", ")" ]
Creates user, encrypts password.
[ "Creates", "user", "encrypts", "password", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L110-L118
zetaops/zengine
zengine/management_commands.py
RunServer.run
def run(self): """ Starts a development server for the zengine application """ print("Development server started on http://%s:%s. \n\nPress Ctrl+C to stop\n" % ( self.manager.args.addr, self.manager.args.port) ) if self.manager.args.server_type == 'falcon': self.run_with_falcon() elif self.manager.args.server_type == 'tornado': self.run_with_tornado()
python
def run(self): """ Starts a development server for the zengine application """ print("Development server started on http://%s:%s. \n\nPress Ctrl+C to stop\n" % ( self.manager.args.addr, self.manager.args.port) ) if self.manager.args.server_type == 'falcon': self.run_with_falcon() elif self.manager.args.server_type == 'tornado': self.run_with_tornado()
[ "def", "run", "(", "self", ")", ":", "print", "(", "\"Development server started on http://%s:%s. \\n\\nPress Ctrl+C to stop\\n\"", "%", "(", "self", ".", "manager", ".", "args", ".", "addr", ",", "self", ".", "manager", ".", "args", ".", "port", ")", ")", "if", "self", ".", "manager", ".", "args", ".", "server_type", "==", "'falcon'", ":", "self", ".", "run_with_falcon", "(", ")", "elif", "self", ".", "manager", ".", "args", ".", "server_type", "==", "'tornado'", ":", "self", ".", "run_with_tornado", "(", ")" ]
Starts a development server for the zengine application
[ "Starts", "a", "development", "server", "for", "the", "zengine", "application" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L139-L150
zetaops/zengine
zengine/management_commands.py
RunServer.run_with_tornado
def run_with_tornado(self): """ runs the tornado/websockets based test server """ from zengine.tornado_server.server import runserver runserver(self.manager.args.addr, int(self.manager.args.port))
python
def run_with_tornado(self): """ runs the tornado/websockets based test server """ from zengine.tornado_server.server import runserver runserver(self.manager.args.addr, int(self.manager.args.port))
[ "def", "run_with_tornado", "(", "self", ")", ":", "from", "zengine", ".", "tornado_server", ".", "server", "import", "runserver", "runserver", "(", "self", ".", "manager", ".", "args", ".", "addr", ",", "int", "(", "self", ".", "manager", ".", "args", ".", "port", ")", ")" ]
runs the tornado/websockets based test server
[ "runs", "the", "tornado", "/", "websockets", "based", "test", "server" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L152-L157
zetaops/zengine
zengine/management_commands.py
RunServer.run_with_falcon
def run_with_falcon(self): """ runs the falcon/http based test server """ from wsgiref import simple_server from zengine.server import app httpd = simple_server.make_server(self.manager.args.addr, int(self.manager.args.port), app) httpd.serve_forever()
python
def run_with_falcon(self): """ runs the falcon/http based test server """ from wsgiref import simple_server from zengine.server import app httpd = simple_server.make_server(self.manager.args.addr, int(self.manager.args.port), app) httpd.serve_forever()
[ "def", "run_with_falcon", "(", "self", ")", ":", "from", "wsgiref", "import", "simple_server", "from", "zengine", ".", "server", "import", "app", "httpd", "=", "simple_server", ".", "make_server", "(", "self", ".", "manager", ".", "args", ".", "addr", ",", "int", "(", "self", ".", "manager", ".", "args", ".", "port", ")", ",", "app", ")", "httpd", ".", "serve_forever", "(", ")" ]
runs the falcon/http based test server
[ "runs", "the", "falcon", "/", "http", "based", "test", "server" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L159-L166
zetaops/zengine
zengine/management_commands.py
RunWorker.run
def run(self): """ Starts a development server for the zengine application """ from zengine.wf_daemon import run_workers, Worker worker_count = int(self.manager.args.workers or 1) if not self.manager.args.daemonize: print("Starting worker(s)") if worker_count > 1 or self.manager.args.autoreload: run_workers(worker_count, self.manager.args.paths.split(' '), self.manager.args.daemonize) else: worker = Worker() worker.run()
python
def run(self): """ Starts a development server for the zengine application """ from zengine.wf_daemon import run_workers, Worker worker_count = int(self.manager.args.workers or 1) if not self.manager.args.daemonize: print("Starting worker(s)") if worker_count > 1 or self.manager.args.autoreload: run_workers(worker_count, self.manager.args.paths.split(' '), self.manager.args.daemonize) else: worker = Worker() worker.run()
[ "def", "run", "(", "self", ")", ":", "from", "zengine", ".", "wf_daemon", "import", "run_workers", ",", "Worker", "worker_count", "=", "int", "(", "self", ".", "manager", ".", "args", ".", "workers", "or", "1", ")", "if", "not", "self", ".", "manager", ".", "args", ".", "daemonize", ":", "print", "(", "\"Starting worker(s)\"", ")", "if", "worker_count", ">", "1", "or", "self", ".", "manager", ".", "args", ".", "autoreload", ":", "run_workers", "(", "worker_count", ",", "self", ".", "manager", ".", "args", ".", "paths", ".", "split", "(", "' '", ")", ",", "self", ".", "manager", ".", "args", ".", "daemonize", ")", "else", ":", "worker", "=", "Worker", "(", ")", "worker", ".", "run", "(", ")" ]
Starts a development server for the zengine application
[ "Starts", "a", "development", "server", "for", "the", "zengine", "application" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L187-L203
zetaops/zengine
zengine/management_commands.py
ExtractTranslations._prepare_domain
def _prepare_domain(mapping): """Prepare a helper dictionary for the domain to temporarily hold some information.""" # Parse the domain-directory mapping try: domain, dir = mapping.split(':') except ValueError: print("Please provide the sources in the form of '<domain>:<directory>'") sys.exit(1) try: default_language = settings.TRANSLATION_DOMAINS[domain] except KeyError: print("Unknown domain {domain}, check the settings file to make sure" " this domain is set in TRANSLATION_DOMAINS".format(domain=domain)) sys.exit(1) # Create a temporary file to hold the `.pot` file for this domain handle, path = tempfile.mkstemp(prefix='zengine_i18n_', suffix='.pot') return (domain, { 'default': default_language, 'pot': path, 'source': dir, })
python
def _prepare_domain(mapping): """Prepare a helper dictionary for the domain to temporarily hold some information.""" # Parse the domain-directory mapping try: domain, dir = mapping.split(':') except ValueError: print("Please provide the sources in the form of '<domain>:<directory>'") sys.exit(1) try: default_language = settings.TRANSLATION_DOMAINS[domain] except KeyError: print("Unknown domain {domain}, check the settings file to make sure" " this domain is set in TRANSLATION_DOMAINS".format(domain=domain)) sys.exit(1) # Create a temporary file to hold the `.pot` file for this domain handle, path = tempfile.mkstemp(prefix='zengine_i18n_', suffix='.pot') return (domain, { 'default': default_language, 'pot': path, 'source': dir, })
[ "def", "_prepare_domain", "(", "mapping", ")", ":", "# Parse the domain-directory mapping", "try", ":", "domain", ",", "dir", "=", "mapping", ".", "split", "(", "':'", ")", "except", "ValueError", ":", "print", "(", "\"Please provide the sources in the form of '<domain>:<directory>'\"", ")", "sys", ".", "exit", "(", "1", ")", "try", ":", "default_language", "=", "settings", ".", "TRANSLATION_DOMAINS", "[", "domain", "]", "except", "KeyError", ":", "print", "(", "\"Unknown domain {domain}, check the settings file to make sure\"", "\" this domain is set in TRANSLATION_DOMAINS\"", ".", "format", "(", "domain", "=", "domain", ")", ")", "sys", ".", "exit", "(", "1", ")", "# Create a temporary file to hold the `.pot` file for this domain", "handle", ",", "path", "=", "tempfile", ".", "mkstemp", "(", "prefix", "=", "'zengine_i18n_'", ",", "suffix", "=", "'.pot'", ")", "return", "(", "domain", ",", "{", "'default'", ":", "default_language", ",", "'pot'", ":", "path", ",", "'source'", ":", "dir", ",", "}", ")" ]
Prepare a helper dictionary for the domain to temporarily hold some information.
[ "Prepare", "a", "helper", "dictionary", "for", "the", "domain", "to", "temporarily", "hold", "some", "information", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L227-L248
zetaops/zengine
zengine/management_commands.py
ExtractTranslations._validate_domains
def _validate_domains(domains): """Check that all domains specified in the settings was provided in the options.""" missing = set(settings.TRANSLATION_DOMAINS.keys()) - set(domains.keys()) if missing: print('The following domains have been set in the configuration, ' 'but their sources were not provided, use the `--source` ' 'option to specify their sources: {domains}'.format(domains=', '.join(missing))) sys.exit(1)
python
def _validate_domains(domains): """Check that all domains specified in the settings was provided in the options.""" missing = set(settings.TRANSLATION_DOMAINS.keys()) - set(domains.keys()) if missing: print('The following domains have been set in the configuration, ' 'but their sources were not provided, use the `--source` ' 'option to specify their sources: {domains}'.format(domains=', '.join(missing))) sys.exit(1)
[ "def", "_validate_domains", "(", "domains", ")", ":", "missing", "=", "set", "(", "settings", ".", "TRANSLATION_DOMAINS", ".", "keys", "(", ")", ")", "-", "set", "(", "domains", ".", "keys", "(", ")", ")", "if", "missing", ":", "print", "(", "'The following domains have been set in the configuration, '", "'but their sources were not provided, use the `--source` '", "'option to specify their sources: {domains}'", ".", "format", "(", "domains", "=", "', '", ".", "join", "(", "missing", ")", ")", ")", "sys", ".", "exit", "(", "1", ")" ]
Check that all domains specified in the settings was provided in the options.
[ "Check", "that", "all", "domains", "specified", "in", "the", "settings", "was", "provided", "in", "the", "options", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L251-L258
zetaops/zengine
zengine/management_commands.py
ExtractTranslations._extract_translations
def _extract_translations(self, domains): """Extract the translations into `.pot` files""" for domain, options in domains.items(): # Create the extractor extractor = babel_frontend.extract_messages() extractor.initialize_options() # The temporary location to write the `.pot` file extractor.output_file = options['pot'] # Add the comments marked with 'tn:' to the translation file for translators to read. Strip the marker. extractor.add_comments = ['tn:'] extractor.strip_comments = True # The directory where the sources for this domain are located extractor.input_paths = [options['source']] # Pass the metadata to the translator extractor.msgid_bugs_address = self.manager.args.contact extractor.copyright_holder = self.manager.args.copyright extractor.version = self.manager.args.version extractor.project = self.manager.args.project extractor.finalize_options() # Add keywords for lazy translation functions, based on their non-lazy variants extractor.keywords.update({ 'gettext_lazy': extractor.keywords['gettext'], 'ngettext_lazy': extractor.keywords['ngettext'], '__': extractor.keywords['gettext'], # double underscore for lazy }) # Do the extraction _run_babel_command(extractor)
python
def _extract_translations(self, domains): """Extract the translations into `.pot` files""" for domain, options in domains.items(): # Create the extractor extractor = babel_frontend.extract_messages() extractor.initialize_options() # The temporary location to write the `.pot` file extractor.output_file = options['pot'] # Add the comments marked with 'tn:' to the translation file for translators to read. Strip the marker. extractor.add_comments = ['tn:'] extractor.strip_comments = True # The directory where the sources for this domain are located extractor.input_paths = [options['source']] # Pass the metadata to the translator extractor.msgid_bugs_address = self.manager.args.contact extractor.copyright_holder = self.manager.args.copyright extractor.version = self.manager.args.version extractor.project = self.manager.args.project extractor.finalize_options() # Add keywords for lazy translation functions, based on their non-lazy variants extractor.keywords.update({ 'gettext_lazy': extractor.keywords['gettext'], 'ngettext_lazy': extractor.keywords['ngettext'], '__': extractor.keywords['gettext'], # double underscore for lazy }) # Do the extraction _run_babel_command(extractor)
[ "def", "_extract_translations", "(", "self", ",", "domains", ")", ":", "for", "domain", ",", "options", "in", "domains", ".", "items", "(", ")", ":", "# Create the extractor", "extractor", "=", "babel_frontend", ".", "extract_messages", "(", ")", "extractor", ".", "initialize_options", "(", ")", "# The temporary location to write the `.pot` file", "extractor", ".", "output_file", "=", "options", "[", "'pot'", "]", "# Add the comments marked with 'tn:' to the translation file for translators to read. Strip the marker.", "extractor", ".", "add_comments", "=", "[", "'tn:'", "]", "extractor", ".", "strip_comments", "=", "True", "# The directory where the sources for this domain are located", "extractor", ".", "input_paths", "=", "[", "options", "[", "'source'", "]", "]", "# Pass the metadata to the translator", "extractor", ".", "msgid_bugs_address", "=", "self", ".", "manager", ".", "args", ".", "contact", "extractor", ".", "copyright_holder", "=", "self", ".", "manager", ".", "args", ".", "copyright", "extractor", ".", "version", "=", "self", ".", "manager", ".", "args", ".", "version", "extractor", ".", "project", "=", "self", ".", "manager", ".", "args", ".", "project", "extractor", ".", "finalize_options", "(", ")", "# Add keywords for lazy translation functions, based on their non-lazy variants", "extractor", ".", "keywords", ".", "update", "(", "{", "'gettext_lazy'", ":", "extractor", ".", "keywords", "[", "'gettext'", "]", ",", "'ngettext_lazy'", ":", "extractor", ".", "keywords", "[", "'ngettext'", "]", ",", "'__'", ":", "extractor", ".", "keywords", "[", "'gettext'", "]", ",", "# double underscore for lazy", "}", ")", "# Do the extraction", "_run_babel_command", "(", "extractor", ")" ]
Extract the translations into `.pot` files
[ "Extract", "the", "translations", "into", ".", "pot", "files" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L260-L286
zetaops/zengine
zengine/management_commands.py
ExtractTranslations._init_update_po_files
def _init_update_po_files(self, domains): """Update or initialize the `.po` translation files""" for language in settings.TRANSLATIONS: for domain, options in domains.items(): if language == options['default']: continue # Default language of the domain doesn't need translations if os.path.isfile(_po_path(language, domain)): # If the translation already exists, update it, keeping the parts already translated self._update_po_file(language, domain, options['pot']) else: # The translation doesn't exist, create a new translation file self._init_po_file(language, domain, options['pot'])
python
def _init_update_po_files(self, domains): """Update or initialize the `.po` translation files""" for language in settings.TRANSLATIONS: for domain, options in domains.items(): if language == options['default']: continue # Default language of the domain doesn't need translations if os.path.isfile(_po_path(language, domain)): # If the translation already exists, update it, keeping the parts already translated self._update_po_file(language, domain, options['pot']) else: # The translation doesn't exist, create a new translation file self._init_po_file(language, domain, options['pot'])
[ "def", "_init_update_po_files", "(", "self", ",", "domains", ")", ":", "for", "language", "in", "settings", ".", "TRANSLATIONS", ":", "for", "domain", ",", "options", "in", "domains", ".", "items", "(", ")", ":", "if", "language", "==", "options", "[", "'default'", "]", ":", "continue", "# Default language of the domain doesn't need translations", "if", "os", ".", "path", ".", "isfile", "(", "_po_path", "(", "language", ",", "domain", ")", ")", ":", "# If the translation already exists, update it, keeping the parts already translated", "self", ".", "_update_po_file", "(", "language", ",", "domain", ",", "options", "[", "'pot'", "]", ")", "else", ":", "# The translation doesn't exist, create a new translation file", "self", ".", "_init_po_file", "(", "language", ",", "domain", ",", "options", "[", "'pot'", "]", ")" ]
Update or initialize the `.po` translation files
[ "Update", "or", "initialize", "the", ".", "po", "translation", "files" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L288-L298
zetaops/zengine
zengine/management_commands.py
ExtractTranslations._cleanup
def _cleanup(self, domains): """Remove the temporary '.pot' files that were created for the domains.""" for option in domains.values(): try: os.remove(option['pot']) except (IOError, OSError): # It is not a problem if we can't actually remove the temporary file pass
python
def _cleanup(self, domains): """Remove the temporary '.pot' files that were created for the domains.""" for option in domains.values(): try: os.remove(option['pot']) except (IOError, OSError): # It is not a problem if we can't actually remove the temporary file pass
[ "def", "_cleanup", "(", "self", ",", "domains", ")", ":", "for", "option", "in", "domains", ".", "values", "(", ")", ":", "try", ":", "os", ".", "remove", "(", "option", "[", "'pot'", "]", ")", "except", "(", "IOError", ",", "OSError", ")", ":", "# It is not a problem if we can't actually remove the temporary file", "pass" ]
Remove the temporary '.pot' files that were created for the domains.
[ "Remove", "the", "temporary", ".", "pot", "files", "that", "were", "created", "for", "the", "domains", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L312-L319
zetaops/zengine
zengine/management_commands.py
LoadDiagrams.run
def run(self): """ read workflows, checks if it's updated, tries to update if there aren't any running instances of that wf """ from zengine.lib.cache import WFSpecNames if self.manager.args.clear: self._clear_models() return if self.manager.args.wf_path: paths = self.get_wf_from_path(self.manager.args.wf_path) else: paths = self.get_workflows() self.count = 0 self.do_with_submit(self.load_diagram, paths, threads=self.manager.args.threads) WFSpecNames().refresh() print("%s BPMN file loaded" % self.count)
python
def run(self): """ read workflows, checks if it's updated, tries to update if there aren't any running instances of that wf """ from zengine.lib.cache import WFSpecNames if self.manager.args.clear: self._clear_models() return if self.manager.args.wf_path: paths = self.get_wf_from_path(self.manager.args.wf_path) else: paths = self.get_workflows() self.count = 0 self.do_with_submit(self.load_diagram, paths, threads=self.manager.args.threads) WFSpecNames().refresh() print("%s BPMN file loaded" % self.count)
[ "def", "run", "(", "self", ")", ":", "from", "zengine", ".", "lib", ".", "cache", "import", "WFSpecNames", "if", "self", ".", "manager", ".", "args", ".", "clear", ":", "self", ".", "_clear_models", "(", ")", "return", "if", "self", ".", "manager", ".", "args", ".", "wf_path", ":", "paths", "=", "self", ".", "get_wf_from_path", "(", "self", ".", "manager", ".", "args", ".", "wf_path", ")", "else", ":", "paths", "=", "self", ".", "get_workflows", "(", ")", "self", ".", "count", "=", "0", "self", ".", "do_with_submit", "(", "self", ".", "load_diagram", ",", "paths", ",", "threads", "=", "self", ".", "manager", ".", "args", ".", "threads", ")", "WFSpecNames", "(", ")", ".", "refresh", "(", ")", "print", "(", "\"%s BPMN file loaded\"", "%", "self", ".", "count", ")" ]
read workflows, checks if it's updated, tries to update if there aren't any running instances of that wf
[ "read", "workflows", "checks", "if", "it", "s", "updated", "tries", "to", "update", "if", "there", "aren", "t", "any", "running", "instances", "of", "that", "wf" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L450-L472
zetaops/zengine
zengine/management_commands.py
LoadDiagrams.get_wf_from_path
def get_wf_from_path(self, path): """ load xml from given path Args: path: diagram path Returns: """ with open(path) as fp: content = fp.read() return [(os.path.basename(os.path.splitext(path)[0]), content), ]
python
def get_wf_from_path(self, path): """ load xml from given path Args: path: diagram path Returns: """ with open(path) as fp: content = fp.read() return [(os.path.basename(os.path.splitext(path)[0]), content), ]
[ "def", "get_wf_from_path", "(", "self", ",", "path", ")", ":", "with", "open", "(", "path", ")", "as", "fp", ":", "content", "=", "fp", ".", "read", "(", ")", "return", "[", "(", "os", ".", "path", ".", "basename", "(", "os", ".", "path", ".", "splitext", "(", "path", ")", "[", "0", "]", ")", ",", "content", ")", ",", "]" ]
load xml from given path Args: path: diagram path Returns:
[ "load", "xml", "from", "given", "path", "Args", ":", "path", ":", "diagram", "path" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L515-L526
zetaops/zengine
zengine/management_commands.py
LoadDiagrams.get_workflows
def get_workflows(self): """ Scans and loads all wf found under WORKFLOW_PACKAGES_PATHS Yields: XML content of diagram file """ for pth in settings.WORKFLOW_PACKAGES_PATHS: for f in glob.glob("%s/*.bpmn" % pth): with open(f) as fp: yield os.path.basename(os.path.splitext(f)[0]), fp.read()
python
def get_workflows(self): """ Scans and loads all wf found under WORKFLOW_PACKAGES_PATHS Yields: XML content of diagram file """ for pth in settings.WORKFLOW_PACKAGES_PATHS: for f in glob.glob("%s/*.bpmn" % pth): with open(f) as fp: yield os.path.basename(os.path.splitext(f)[0]), fp.read()
[ "def", "get_workflows", "(", "self", ")", ":", "for", "pth", "in", "settings", ".", "WORKFLOW_PACKAGES_PATHS", ":", "for", "f", "in", "glob", ".", "glob", "(", "\"%s/*.bpmn\"", "%", "pth", ")", ":", "with", "open", "(", "f", ")", "as", "fp", ":", "yield", "os", ".", "path", ".", "basename", "(", "os", ".", "path", ".", "splitext", "(", "f", ")", "[", "0", "]", ")", ",", "fp", ".", "read", "(", ")" ]
Scans and loads all wf found under WORKFLOW_PACKAGES_PATHS Yields: XML content of diagram file
[ "Scans", "and", "loads", "all", "wf", "found", "under", "WORKFLOW_PACKAGES_PATHS" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L528-L538
zetaops/zengine
zengine/management_commands.py
CheckList.check_migration_and_solr
def check_migration_and_solr(self): """ The model or models are checked for migrations that need to be done. Solr is also checked. """ from pyoko.db.schema_update import SchemaUpdater from socket import error as socket_error from pyoko.conf import settings from importlib import import_module import_module(settings.MODELS_MODULE) registry = import_module('pyoko.model').model_registry models = [model for model in registry.get_base_models()] try: print(__(u"Checking migration and solr ...")) updater = SchemaUpdater(models, 1, False) updater.run(check_only=True) except socket_error as e: print(__(u"{0}Error not connected, open redis and rabbitmq{1}").format(CheckList.FAIL, CheckList.ENDC))
python
def check_migration_and_solr(self): """ The model or models are checked for migrations that need to be done. Solr is also checked. """ from pyoko.db.schema_update import SchemaUpdater from socket import error as socket_error from pyoko.conf import settings from importlib import import_module import_module(settings.MODELS_MODULE) registry = import_module('pyoko.model').model_registry models = [model for model in registry.get_base_models()] try: print(__(u"Checking migration and solr ...")) updater = SchemaUpdater(models, 1, False) updater.run(check_only=True) except socket_error as e: print(__(u"{0}Error not connected, open redis and rabbitmq{1}").format(CheckList.FAIL, CheckList.ENDC))
[ "def", "check_migration_and_solr", "(", "self", ")", ":", "from", "pyoko", ".", "db", ".", "schema_update", "import", "SchemaUpdater", "from", "socket", "import", "error", "as", "socket_error", "from", "pyoko", ".", "conf", "import", "settings", "from", "importlib", "import", "import_module", "import_module", "(", "settings", ".", "MODELS_MODULE", ")", "registry", "=", "import_module", "(", "'pyoko.model'", ")", ".", "model_registry", "models", "=", "[", "model", "for", "model", "in", "registry", ".", "get_base_models", "(", ")", "]", "try", ":", "print", "(", "__", "(", "u\"Checking migration and solr ...\"", ")", ")", "updater", "=", "SchemaUpdater", "(", "models", ",", "1", ",", "False", ")", "updater", ".", "run", "(", "check_only", "=", "True", ")", "except", "socket_error", "as", "e", ":", "print", "(", "__", "(", "u\"{0}Error not connected, open redis and rabbitmq{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ")" ]
The model or models are checked for migrations that need to be done. Solr is also checked.
[ "The", "model", "or", "models", "are", "checked", "for", "migrations", "that", "need", "to", "be", "done", ".", "Solr", "is", "also", "checked", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L563-L583
zetaops/zengine
zengine/management_commands.py
CheckList.check_redis
def check_redis(): """ Redis checks the connection It displays on the screen whether or not you have a connection. """ from pyoko.db.connection import cache from redis.exceptions import ConnectionError try: cache.ping() print(CheckList.OKGREEN + "{0}Redis is working{1}" + CheckList.ENDC) except ConnectionError as e: print(__(u"{0}Redis is not working{1} ").format(CheckList.FAIL, CheckList.ENDC), e.message)
python
def check_redis(): """ Redis checks the connection It displays on the screen whether or not you have a connection. """ from pyoko.db.connection import cache from redis.exceptions import ConnectionError try: cache.ping() print(CheckList.OKGREEN + "{0}Redis is working{1}" + CheckList.ENDC) except ConnectionError as e: print(__(u"{0}Redis is not working{1} ").format(CheckList.FAIL, CheckList.ENDC), e.message)
[ "def", "check_redis", "(", ")", ":", "from", "pyoko", ".", "db", ".", "connection", "import", "cache", "from", "redis", ".", "exceptions", "import", "ConnectionError", "try", ":", "cache", ".", "ping", "(", ")", "print", "(", "CheckList", ".", "OKGREEN", "+", "\"{0}Redis is working{1}\"", "+", "CheckList", ".", "ENDC", ")", "except", "ConnectionError", "as", "e", ":", "print", "(", "__", "(", "u\"{0}Redis is not working{1} \"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ",", "e", ".", "message", ")" ]
Redis checks the connection It displays on the screen whether or not you have a connection.
[ "Redis", "checks", "the", "connection", "It", "displays", "on", "the", "screen", "whether", "or", "not", "you", "have", "a", "connection", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L586-L599
zetaops/zengine
zengine/management_commands.py
CheckList.check_riak
def check_riak(): """ Riak checks the connection It displays on the screen whether or not you have a connection. """ from pyoko.db.connection import client from socket import error as socket_error try: if client.ping(): print(__(u"{0}Riak is working{1}").format(CheckList.OKGREEN, CheckList.ENDC)) else: print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL, CheckList.ENDC)) except socket_error as e: print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL, CheckList.ENDC), e.message)
python
def check_riak(): """ Riak checks the connection It displays on the screen whether or not you have a connection. """ from pyoko.db.connection import client from socket import error as socket_error try: if client.ping(): print(__(u"{0}Riak is working{1}").format(CheckList.OKGREEN, CheckList.ENDC)) else: print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL, CheckList.ENDC)) except socket_error as e: print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL, CheckList.ENDC), e.message)
[ "def", "check_riak", "(", ")", ":", "from", "pyoko", ".", "db", ".", "connection", "import", "client", "from", "socket", "import", "error", "as", "socket_error", "try", ":", "if", "client", ".", "ping", "(", ")", ":", "print", "(", "__", "(", "u\"{0}Riak is working{1}\"", ")", ".", "format", "(", "CheckList", ".", "OKGREEN", ",", "CheckList", ".", "ENDC", ")", ")", "else", ":", "print", "(", "__", "(", "u\"{0}Riak is not working{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ")", "except", "socket_error", "as", "e", ":", "print", "(", "__", "(", "u\"{0}Riak is not working{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ",", "e", ".", "message", ")" ]
Riak checks the connection It displays on the screen whether or not you have a connection.
[ "Riak", "checks", "the", "connection", "It", "displays", "on", "the", "screen", "whether", "or", "not", "you", "have", "a", "connection", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L602-L617
zetaops/zengine
zengine/management_commands.py
CheckList.check_mq_connection
def check_mq_connection(self): """ RabbitMQ checks the connection It displays on the screen whether or not you have a connection. """ import pika from zengine.client_queue import BLOCKING_MQ_PARAMS from pika.exceptions import ProbableAuthenticationError, ConnectionClosed try: connection = pika.BlockingConnection(BLOCKING_MQ_PARAMS) channel = connection.channel() if channel.is_open: print(__(u"{0}RabbitMQ is working{1}").format(CheckList.OKGREEN, CheckList.ENDC)) elif self.channel.is_closed or self.channel.is_closing: print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC)) except ConnectionClosed as e: print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC), e) except ProbableAuthenticationError as e: print(__(u"{0}RabbitMQ username and password wrong{1}").format(CheckList.FAIL, CheckList.ENDC))
python
def check_mq_connection(self): """ RabbitMQ checks the connection It displays on the screen whether or not you have a connection. """ import pika from zengine.client_queue import BLOCKING_MQ_PARAMS from pika.exceptions import ProbableAuthenticationError, ConnectionClosed try: connection = pika.BlockingConnection(BLOCKING_MQ_PARAMS) channel = connection.channel() if channel.is_open: print(__(u"{0}RabbitMQ is working{1}").format(CheckList.OKGREEN, CheckList.ENDC)) elif self.channel.is_closed or self.channel.is_closing: print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC)) except ConnectionClosed as e: print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC), e) except ProbableAuthenticationError as e: print(__(u"{0}RabbitMQ username and password wrong{1}").format(CheckList.FAIL, CheckList.ENDC))
[ "def", "check_mq_connection", "(", "self", ")", ":", "import", "pika", "from", "zengine", ".", "client_queue", "import", "BLOCKING_MQ_PARAMS", "from", "pika", ".", "exceptions", "import", "ProbableAuthenticationError", ",", "ConnectionClosed", "try", ":", "connection", "=", "pika", ".", "BlockingConnection", "(", "BLOCKING_MQ_PARAMS", ")", "channel", "=", "connection", ".", "channel", "(", ")", "if", "channel", ".", "is_open", ":", "print", "(", "__", "(", "u\"{0}RabbitMQ is working{1}\"", ")", ".", "format", "(", "CheckList", ".", "OKGREEN", ",", "CheckList", ".", "ENDC", ")", ")", "elif", "self", ".", "channel", ".", "is_closed", "or", "self", ".", "channel", ".", "is_closing", ":", "print", "(", "__", "(", "u\"{0}RabbitMQ is not working!{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ")", "except", "ConnectionClosed", "as", "e", ":", "print", "(", "__", "(", "u\"{0}RabbitMQ is not working!{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ",", "e", ")", "except", "ProbableAuthenticationError", "as", "e", ":", "print", "(", "__", "(", "u\"{0}RabbitMQ username and password wrong{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ")" ]
RabbitMQ checks the connection It displays on the screen whether or not you have a connection.
[ "RabbitMQ", "checks", "the", "connection", "It", "displays", "on", "the", "screen", "whether", "or", "not", "you", "have", "a", "connection", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L619-L639
zetaops/zengine
zengine/management_commands.py
CheckList.check_encoding_and_env
def check_encoding_and_env(): """ It brings the environment variables to the screen. The user checks to see if they are using the correct variables. """ import sys import os if sys.getfilesystemencoding() in ['utf-8', 'UTF-8']: print(__(u"{0}File system encoding correct{1}").format(CheckList.OKGREEN, CheckList.ENDC)) else: print(__(u"{0}File system encoding wrong!!{1}").format(CheckList.FAIL, CheckList.ENDC)) check_env_list = ['RIAK_PROTOCOL', 'RIAK_SERVER', 'RIAK_PORT', 'REDIS_SERVER', 'DEFAULT_BUCKET_TYPE', 'PYOKO_SETTINGS', 'MQ_HOST', 'MQ_PORT', 'MQ_USER', 'MQ_VHOST', ] env = os.environ for k, v in env.items(): if k in check_env_list: print(__(u"{0}{1} : {2}{3}").format(CheckList.BOLD, k, v, CheckList.ENDC))
python
def check_encoding_and_env(): """ It brings the environment variables to the screen. The user checks to see if they are using the correct variables. """ import sys import os if sys.getfilesystemencoding() in ['utf-8', 'UTF-8']: print(__(u"{0}File system encoding correct{1}").format(CheckList.OKGREEN, CheckList.ENDC)) else: print(__(u"{0}File system encoding wrong!!{1}").format(CheckList.FAIL, CheckList.ENDC)) check_env_list = ['RIAK_PROTOCOL', 'RIAK_SERVER', 'RIAK_PORT', 'REDIS_SERVER', 'DEFAULT_BUCKET_TYPE', 'PYOKO_SETTINGS', 'MQ_HOST', 'MQ_PORT', 'MQ_USER', 'MQ_VHOST', ] env = os.environ for k, v in env.items(): if k in check_env_list: print(__(u"{0}{1} : {2}{3}").format(CheckList.BOLD, k, v, CheckList.ENDC))
[ "def", "check_encoding_and_env", "(", ")", ":", "import", "sys", "import", "os", "if", "sys", ".", "getfilesystemencoding", "(", ")", "in", "[", "'utf-8'", ",", "'UTF-8'", "]", ":", "print", "(", "__", "(", "u\"{0}File system encoding correct{1}\"", ")", ".", "format", "(", "CheckList", ".", "OKGREEN", ",", "CheckList", ".", "ENDC", ")", ")", "else", ":", "print", "(", "__", "(", "u\"{0}File system encoding wrong!!{1}\"", ")", ".", "format", "(", "CheckList", ".", "FAIL", ",", "CheckList", ".", "ENDC", ")", ")", "check_env_list", "=", "[", "'RIAK_PROTOCOL'", ",", "'RIAK_SERVER'", ",", "'RIAK_PORT'", ",", "'REDIS_SERVER'", ",", "'DEFAULT_BUCKET_TYPE'", ",", "'PYOKO_SETTINGS'", ",", "'MQ_HOST'", ",", "'MQ_PORT'", ",", "'MQ_USER'", ",", "'MQ_VHOST'", ",", "]", "env", "=", "os", ".", "environ", "for", "k", ",", "v", "in", "env", ".", "items", "(", ")", ":", "if", "k", "in", "check_env_list", ":", "print", "(", "__", "(", "u\"{0}{1} : {2}{3}\"", ")", ".", "format", "(", "CheckList", ".", "BOLD", ",", "k", ",", "v", ",", "CheckList", ".", "ENDC", ")", ")" ]
It brings the environment variables to the screen. The user checks to see if they are using the correct variables.
[ "It", "brings", "the", "environment", "variables", "to", "the", "screen", ".", "The", "user", "checks", "to", "see", "if", "they", "are", "using", "the", "correct", "variables", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L642-L662
LordDarkula/chess_py
chess_py/game/game_state.py
no_moves
def no_moves(position): """ Finds if the game is over. :type: position: Board :rtype: bool """ return position.no_moves(color.white) \ or position.no_moves(color.black)
python
def no_moves(position): """ Finds if the game is over. :type: position: Board :rtype: bool """ return position.no_moves(color.white) \ or position.no_moves(color.black)
[ "def", "no_moves", "(", "position", ")", ":", "return", "position", ".", "no_moves", "(", "color", ".", "white", ")", "or", "position", ".", "no_moves", "(", "color", ".", "black", ")" ]
Finds if the game is over. :type: position: Board :rtype: bool
[ "Finds", "if", "the", "game", "is", "over", "." ]
train
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game_state.py#L13-L21
LordDarkula/chess_py
chess_py/game/game_state.py
is_checkmate
def is_checkmate(position, input_color): """ Finds if particular King is checkmated. :type: position: Board :type: input_color: Color :rtype: bool """ return position.no_moves(input_color) and \ position.get_king(input_color).in_check(position)
python
def is_checkmate(position, input_color): """ Finds if particular King is checkmated. :type: position: Board :type: input_color: Color :rtype: bool """ return position.no_moves(input_color) and \ position.get_king(input_color).in_check(position)
[ "def", "is_checkmate", "(", "position", ",", "input_color", ")", ":", "return", "position", ".", "no_moves", "(", "input_color", ")", "and", "position", ".", "get_king", "(", "input_color", ")", ".", "in_check", "(", "position", ")" ]
Finds if particular King is checkmated. :type: position: Board :type: input_color: Color :rtype: bool
[ "Finds", "if", "particular", "King", "is", "checkmated", "." ]
train
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game_state.py#L24-L33
zetaops/zengine
zengine/messaging/views.py
_paginate
def _paginate(self, current_page, query_set, per_page=10): """ Handles pagination of object listings. Args: current_page int: Current page number query_set (:class:`QuerySet<pyoko:pyoko.db.queryset.QuerySet>`): Object listing queryset. per_page int: Objects per page. Returns: QuerySet object, pagination data dict as a tuple """ total_objects = query_set.count() total_pages = int(total_objects / per_page or 1) # add orphans to last page current_per_page = per_page + ( total_objects % per_page if current_page == total_pages else 0) pagination_data = dict(page=current_page, total_pages=total_pages, total_objects=total_objects, per_page=current_per_page) query_set = query_set.set_params(rows=current_per_page, start=(current_page - 1) * per_page) return query_set, pagination_data
python
def _paginate(self, current_page, query_set, per_page=10): """ Handles pagination of object listings. Args: current_page int: Current page number query_set (:class:`QuerySet<pyoko:pyoko.db.queryset.QuerySet>`): Object listing queryset. per_page int: Objects per page. Returns: QuerySet object, pagination data dict as a tuple """ total_objects = query_set.count() total_pages = int(total_objects / per_page or 1) # add orphans to last page current_per_page = per_page + ( total_objects % per_page if current_page == total_pages else 0) pagination_data = dict(page=current_page, total_pages=total_pages, total_objects=total_objects, per_page=current_per_page) query_set = query_set.set_params(rows=current_per_page, start=(current_page - 1) * per_page) return query_set, pagination_data
[ "def", "_paginate", "(", "self", ",", "current_page", ",", "query_set", ",", "per_page", "=", "10", ")", ":", "total_objects", "=", "query_set", ".", "count", "(", ")", "total_pages", "=", "int", "(", "total_objects", "/", "per_page", "or", "1", ")", "# add orphans to last page", "current_per_page", "=", "per_page", "+", "(", "total_objects", "%", "per_page", "if", "current_page", "==", "total_pages", "else", "0", ")", "pagination_data", "=", "dict", "(", "page", "=", "current_page", ",", "total_pages", "=", "total_pages", ",", "total_objects", "=", "total_objects", ",", "per_page", "=", "current_per_page", ")", "query_set", "=", "query_set", ".", "set_params", "(", "rows", "=", "current_per_page", ",", "start", "=", "(", "current_page", "-", "1", ")", "*", "per_page", ")", "return", "query_set", ",", "pagination_data" ]
Handles pagination of object listings. Args: current_page int: Current page number query_set (:class:`QuerySet<pyoko:pyoko.db.queryset.QuerySet>`): Object listing queryset. per_page int: Objects per page. Returns: QuerySet object, pagination data dict as a tuple
[ "Handles", "pagination", "of", "object", "listings", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L72-L97
zetaops/zengine
zengine/messaging/views.py
create_message
def create_message(current): """ Creates a message for the given channel. .. code-block:: python # request: { 'view':'_zops_create_message', 'message': { 'channel': key, # of channel 'body': string, # message text., 'type': int, # zengine.messaging.model.MSG_TYPES, 'attachments': [{ 'description': string, # can be blank, 'name': string, # file name with extension, 'content': string, # base64 encoded file content }]} # response: { 'status': 'Created', 'code': 201, 'msg_key': key, # key of the message object, } """ msg = current.input['message'] msg_obj = Channel.add_message(msg['channel'], body=msg['body'], typ=msg['type'], sender=current.user, title=msg['title'], receiver=msg['receiver'] or None) current.output = { 'msg_key': msg_obj.key, 'status': 'Created', 'code': 201 } if 'attachment' in msg: for atch in msg['attachments']: typ = current._dedect_file_type(atch['name'], atch['content']) Attachment(channel_id=msg['channel'], msg=msg_obj, name=atch['name'], file=atch['content'], description=atch['description'], typ=typ).save()
python
def create_message(current): """ Creates a message for the given channel. .. code-block:: python # request: { 'view':'_zops_create_message', 'message': { 'channel': key, # of channel 'body': string, # message text., 'type': int, # zengine.messaging.model.MSG_TYPES, 'attachments': [{ 'description': string, # can be blank, 'name': string, # file name with extension, 'content': string, # base64 encoded file content }]} # response: { 'status': 'Created', 'code': 201, 'msg_key': key, # key of the message object, } """ msg = current.input['message'] msg_obj = Channel.add_message(msg['channel'], body=msg['body'], typ=msg['type'], sender=current.user, title=msg['title'], receiver=msg['receiver'] or None) current.output = { 'msg_key': msg_obj.key, 'status': 'Created', 'code': 201 } if 'attachment' in msg: for atch in msg['attachments']: typ = current._dedect_file_type(atch['name'], atch['content']) Attachment(channel_id=msg['channel'], msg=msg_obj, name=atch['name'], file=atch['content'], description=atch['description'], typ=typ).save()
[ "def", "create_message", "(", "current", ")", ":", "msg", "=", "current", ".", "input", "[", "'message'", "]", "msg_obj", "=", "Channel", ".", "add_message", "(", "msg", "[", "'channel'", "]", ",", "body", "=", "msg", "[", "'body'", "]", ",", "typ", "=", "msg", "[", "'type'", "]", ",", "sender", "=", "current", ".", "user", ",", "title", "=", "msg", "[", "'title'", "]", ",", "receiver", "=", "msg", "[", "'receiver'", "]", "or", "None", ")", "current", ".", "output", "=", "{", "'msg_key'", ":", "msg_obj", ".", "key", ",", "'status'", ":", "'Created'", ",", "'code'", ":", "201", "}", "if", "'attachment'", "in", "msg", ":", "for", "atch", "in", "msg", "[", "'attachments'", "]", ":", "typ", "=", "current", ".", "_dedect_file_type", "(", "atch", "[", "'name'", "]", ",", "atch", "[", "'content'", "]", ")", "Attachment", "(", "channel_id", "=", "msg", "[", "'channel'", "]", ",", "msg", "=", "msg_obj", ",", "name", "=", "atch", "[", "'name'", "]", ",", "file", "=", "atch", "[", "'content'", "]", ",", "description", "=", "atch", "[", "'description'", "]", ",", "typ", "=", "typ", ")", ".", "save", "(", ")" ]
Creates a message for the given channel. .. code-block:: python # request: { 'view':'_zops_create_message', 'message': { 'channel': key, # of channel 'body': string, # message text., 'type': int, # zengine.messaging.model.MSG_TYPES, 'attachments': [{ 'description': string, # can be blank, 'name': string, # file name with extension, 'content': string, # base64 encoded file content }]} # response: { 'status': 'Created', 'code': 201, 'msg_key': key, # key of the message object, }
[ "Creates", "a", "message", "for", "the", "given", "channel", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L100-L139
zetaops/zengine
zengine/messaging/views.py
show_channel
def show_channel(current, waited=False): """ Initial display of channel content. Returns channel description, members, no of members, last 20 messages etc. .. code-block:: python # request: { 'view':'_zops_show_channel', 'key': key, } # response: { 'channel_key': key, 'description': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'name': string, 'last_messages': [MSG_DICT] 'status': 'OK', 'code': 200 } """ ch = Channel(current).objects.get(current.input['key']) sbs = ch.get_subscription_for_user(current.user_id) current.output = {'key': current.input['key'], 'description': ch.description, 'name': sbs.name, 'actions': sbs.get_actions(), 'avatar_url': ch.get_avatar(current.user), 'no_of_members': len(ch.subscriber_set), 'member_list': [{'name': sb.user.full_name, 'is_online': sb.user.is_online(), 'avatar_url': sb.user.get_avatar_url() } for sb in ch.subscriber_set.objects.all()], 'last_messages': [], 'status': 'OK', 'code': 200 } for msg in ch.get_last_messages(): current.output['last_messages'].insert(0, msg.serialize(current.user))
python
def show_channel(current, waited=False): """ Initial display of channel content. Returns channel description, members, no of members, last 20 messages etc. .. code-block:: python # request: { 'view':'_zops_show_channel', 'key': key, } # response: { 'channel_key': key, 'description': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'name': string, 'last_messages': [MSG_DICT] 'status': 'OK', 'code': 200 } """ ch = Channel(current).objects.get(current.input['key']) sbs = ch.get_subscription_for_user(current.user_id) current.output = {'key': current.input['key'], 'description': ch.description, 'name': sbs.name, 'actions': sbs.get_actions(), 'avatar_url': ch.get_avatar(current.user), 'no_of_members': len(ch.subscriber_set), 'member_list': [{'name': sb.user.full_name, 'is_online': sb.user.is_online(), 'avatar_url': sb.user.get_avatar_url() } for sb in ch.subscriber_set.objects.all()], 'last_messages': [], 'status': 'OK', 'code': 200 } for msg in ch.get_last_messages(): current.output['last_messages'].insert(0, msg.serialize(current.user))
[ "def", "show_channel", "(", "current", ",", "waited", "=", "False", ")", ":", "ch", "=", "Channel", "(", "current", ")", ".", "objects", ".", "get", "(", "current", ".", "input", "[", "'key'", "]", ")", "sbs", "=", "ch", ".", "get_subscription_for_user", "(", "current", ".", "user_id", ")", "current", ".", "output", "=", "{", "'key'", ":", "current", ".", "input", "[", "'key'", "]", ",", "'description'", ":", "ch", ".", "description", ",", "'name'", ":", "sbs", ".", "name", ",", "'actions'", ":", "sbs", ".", "get_actions", "(", ")", ",", "'avatar_url'", ":", "ch", ".", "get_avatar", "(", "current", ".", "user", ")", ",", "'no_of_members'", ":", "len", "(", "ch", ".", "subscriber_set", ")", ",", "'member_list'", ":", "[", "{", "'name'", ":", "sb", ".", "user", ".", "full_name", ",", "'is_online'", ":", "sb", ".", "user", ".", "is_online", "(", ")", ",", "'avatar_url'", ":", "sb", ".", "user", ".", "get_avatar_url", "(", ")", "}", "for", "sb", "in", "ch", ".", "subscriber_set", ".", "objects", ".", "all", "(", ")", "]", ",", "'last_messages'", ":", "[", "]", ",", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}", "for", "msg", "in", "ch", ".", "get_last_messages", "(", ")", ":", "current", ".", "output", "[", "'last_messages'", "]", ".", "insert", "(", "0", ",", "msg", ".", "serialize", "(", "current", ".", "user", ")", ")" ]
Initial display of channel content. Returns channel description, members, no of members, last 20 messages etc. .. code-block:: python # request: { 'view':'_zops_show_channel', 'key': key, } # response: { 'channel_key': key, 'description': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'name': string, 'last_messages': [MSG_DICT] 'status': 'OK', 'code': 200 }
[ "Initial", "display", "of", "channel", "content", ".", "Returns", "channel", "description", "members", "no", "of", "members", "last", "20", "messages", "etc", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L142-L189
zetaops/zengine
zengine/messaging/views.py
channel_history
def channel_history(current): """ Get old messages for a channel. 20 messages per request .. code-block:: python # request: { 'view':'_zops_channel_history, 'channel_key': key, 'timestamp': datetime, # timestamp data of oldest shown message } # response: { 'messages': [MSG_DICT, ], 'status': 'OK', 'code': 200 } """ current.output = { 'status': 'OK', 'code': 201, 'messages': [] } for msg in list(Message.objects.filter(channel_id=current.input['channel_key'], updated_at__lte=current.input['timestamp'])[:20]): current.output['messages'].insert(0, msg.serialize(current.user)) # FIXME: looks like pyoko's __lt is broken # TODO: convert lte to lt and remove this block, when __lt filter fixed if current.output['messages']: current.output['messages'].pop(-1)
python
def channel_history(current): """ Get old messages for a channel. 20 messages per request .. code-block:: python # request: { 'view':'_zops_channel_history, 'channel_key': key, 'timestamp': datetime, # timestamp data of oldest shown message } # response: { 'messages': [MSG_DICT, ], 'status': 'OK', 'code': 200 } """ current.output = { 'status': 'OK', 'code': 201, 'messages': [] } for msg in list(Message.objects.filter(channel_id=current.input['channel_key'], updated_at__lte=current.input['timestamp'])[:20]): current.output['messages'].insert(0, msg.serialize(current.user)) # FIXME: looks like pyoko's __lt is broken # TODO: convert lte to lt and remove this block, when __lt filter fixed if current.output['messages']: current.output['messages'].pop(-1)
[ "def", "channel_history", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "201", ",", "'messages'", ":", "[", "]", "}", "for", "msg", "in", "list", "(", "Message", ".", "objects", ".", "filter", "(", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ",", "updated_at__lte", "=", "current", ".", "input", "[", "'timestamp'", "]", ")", "[", ":", "20", "]", ")", ":", "current", ".", "output", "[", "'messages'", "]", ".", "insert", "(", "0", ",", "msg", ".", "serialize", "(", "current", ".", "user", ")", ")", "# FIXME: looks like pyoko's __lt is broken", "# TODO: convert lte to lt and remove this block, when __lt filter fixed", "if", "current", ".", "output", "[", "'messages'", "]", ":", "current", ".", "output", "[", "'messages'", "]", ".", "pop", "(", "-", "1", ")" ]
Get old messages for a channel. 20 messages per request .. code-block:: python # request: { 'view':'_zops_channel_history, 'channel_key': key, 'timestamp': datetime, # timestamp data of oldest shown message } # response: { 'messages': [MSG_DICT, ], 'status': 'OK', 'code': 200 }
[ "Get", "old", "messages", "for", "a", "channel", ".", "20", "messages", "per", "request" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L192-L224
zetaops/zengine
zengine/messaging/views.py
report_last_seen_message
def report_last_seen_message(current): """ Push timestamp of latest message of an ACTIVE channel. This view should be called with timestamp of latest message; - When user opens (clicks on) a channel. - Periodically (eg: setInterval for 15secs) while user staying in a channel. .. code-block:: python # request: { 'view':'_zops_last_seen_msg', 'channel_key': key, 'key': key, 'timestamp': datetime, } # response: { 'status': 'OK', 'code': 200, } """ sbs = Subscriber(current).objects.filter(channel_id=current.input['channel_key'], user_id=current.user_id)[0] sbs.last_seen_msg_time = current.input['timestamp'] sbs.save() current.output = { 'status': 'OK', 'code': 200}
python
def report_last_seen_message(current): """ Push timestamp of latest message of an ACTIVE channel. This view should be called with timestamp of latest message; - When user opens (clicks on) a channel. - Periodically (eg: setInterval for 15secs) while user staying in a channel. .. code-block:: python # request: { 'view':'_zops_last_seen_msg', 'channel_key': key, 'key': key, 'timestamp': datetime, } # response: { 'status': 'OK', 'code': 200, } """ sbs = Subscriber(current).objects.filter(channel_id=current.input['channel_key'], user_id=current.user_id)[0] sbs.last_seen_msg_time = current.input['timestamp'] sbs.save() current.output = { 'status': 'OK', 'code': 200}
[ "def", "report_last_seen_message", "(", "current", ")", ":", "sbs", "=", "Subscriber", "(", "current", ")", ".", "objects", ".", "filter", "(", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ",", "user_id", "=", "current", ".", "user_id", ")", "[", "0", "]", "sbs", ".", "last_seen_msg_time", "=", "current", ".", "input", "[", "'timestamp'", "]", "sbs", ".", "save", "(", ")", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}" ]
Push timestamp of latest message of an ACTIVE channel. This view should be called with timestamp of latest message; - When user opens (clicks on) a channel. - Periodically (eg: setInterval for 15secs) while user staying in a channel. .. code-block:: python # request: { 'view':'_zops_last_seen_msg', 'channel_key': key, 'key': key, 'timestamp': datetime, } # response: { 'status': 'OK', 'code': 200, }
[ "Push", "timestamp", "of", "latest", "message", "of", "an", "ACTIVE", "channel", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L227-L258
zetaops/zengine
zengine/messaging/views.py
list_channels
def list_channels(current): """ List channel memberships of current user .. code-block:: python # request: { 'view':'_zops_list_channels', } # response: { 'channels': [ {'name': string, # name of channel 'key': key, # key of channel 'unread': int, # unread message count 'type': int, # channel type, # 15: public channels (chat room/broadcast channel distinction comes from "read_only" flag) # 10: direct channels # 5: one and only private channel which is "Notifications" 'read_only': boolean, # true if this is a read-only subscription to a broadcast channel # false if it's a public chat room 'actions':[('action name', 'view name'),] },] } """ current.output = { 'status': 'OK', 'code': 200, 'channels': []} for sbs in current.user.subscriptions.objects.filter(is_visible=True): try: current.output['channels'].append(sbs.get_channel_listing()) except ObjectDoesNotExist: # FIXME: This should not happen, log.exception("UNPAIRED DIRECT EXCHANGES!!!!") sbs.delete()
python
def list_channels(current): """ List channel memberships of current user .. code-block:: python # request: { 'view':'_zops_list_channels', } # response: { 'channels': [ {'name': string, # name of channel 'key': key, # key of channel 'unread': int, # unread message count 'type': int, # channel type, # 15: public channels (chat room/broadcast channel distinction comes from "read_only" flag) # 10: direct channels # 5: one and only private channel which is "Notifications" 'read_only': boolean, # true if this is a read-only subscription to a broadcast channel # false if it's a public chat room 'actions':[('action name', 'view name'),] },] } """ current.output = { 'status': 'OK', 'code': 200, 'channels': []} for sbs in current.user.subscriptions.objects.filter(is_visible=True): try: current.output['channels'].append(sbs.get_channel_listing()) except ObjectDoesNotExist: # FIXME: This should not happen, log.exception("UNPAIRED DIRECT EXCHANGES!!!!") sbs.delete()
[ "def", "list_channels", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", ",", "'channels'", ":", "[", "]", "}", "for", "sbs", "in", "current", ".", "user", ".", "subscriptions", ".", "objects", ".", "filter", "(", "is_visible", "=", "True", ")", ":", "try", ":", "current", ".", "output", "[", "'channels'", "]", ".", "append", "(", "sbs", ".", "get_channel_listing", "(", ")", ")", "except", "ObjectDoesNotExist", ":", "# FIXME: This should not happen,", "log", ".", "exception", "(", "\"UNPAIRED DIRECT EXCHANGES!!!!\"", ")", "sbs", ".", "delete", "(", ")" ]
List channel memberships of current user .. code-block:: python # request: { 'view':'_zops_list_channels', } # response: { 'channels': [ {'name': string, # name of channel 'key': key, # key of channel 'unread': int, # unread message count 'type': int, # channel type, # 15: public channels (chat room/broadcast channel distinction comes from "read_only" flag) # 10: direct channels # 5: one and only private channel which is "Notifications" 'read_only': boolean, # true if this is a read-only subscription to a broadcast channel # false if it's a public chat room 'actions':[('action name', 'view name'),] },] }
[ "List", "channel", "memberships", "of", "current", "user" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L261-L302
zetaops/zengine
zengine/messaging/views.py
unread_count
def unread_count(current): """ Number of unread messages for current user .. code-block:: python # request: { 'view':'_zops_unread_count', } # response: { 'status': 'OK', 'code': 200, 'notifications': int, 'messages': int, } """ unread_ntf = 0 unread_msg = 0 for sbs in current.user.subscriptions.objects.filter(is_visible=True): try: if sbs.channel.key == current.user.prv_exchange: unread_ntf += sbs.unread_count() else: unread_msg += sbs.unread_count() except ObjectDoesNotExist: # FIXME: This should not happen, log.exception("MULTIPLE PRV EXCHANGES!!!!") sbs.delete() current.output = { 'status': 'OK', 'code': 200, 'notifications': unread_ntf, 'messages': unread_msg }
python
def unread_count(current): """ Number of unread messages for current user .. code-block:: python # request: { 'view':'_zops_unread_count', } # response: { 'status': 'OK', 'code': 200, 'notifications': int, 'messages': int, } """ unread_ntf = 0 unread_msg = 0 for sbs in current.user.subscriptions.objects.filter(is_visible=True): try: if sbs.channel.key == current.user.prv_exchange: unread_ntf += sbs.unread_count() else: unread_msg += sbs.unread_count() except ObjectDoesNotExist: # FIXME: This should not happen, log.exception("MULTIPLE PRV EXCHANGES!!!!") sbs.delete() current.output = { 'status': 'OK', 'code': 200, 'notifications': unread_ntf, 'messages': unread_msg }
[ "def", "unread_count", "(", "current", ")", ":", "unread_ntf", "=", "0", "unread_msg", "=", "0", "for", "sbs", "in", "current", ".", "user", ".", "subscriptions", ".", "objects", ".", "filter", "(", "is_visible", "=", "True", ")", ":", "try", ":", "if", "sbs", ".", "channel", ".", "key", "==", "current", ".", "user", ".", "prv_exchange", ":", "unread_ntf", "+=", "sbs", ".", "unread_count", "(", ")", "else", ":", "unread_msg", "+=", "sbs", ".", "unread_count", "(", ")", "except", "ObjectDoesNotExist", ":", "# FIXME: This should not happen,", "log", ".", "exception", "(", "\"MULTIPLE PRV EXCHANGES!!!!\"", ")", "sbs", ".", "delete", "(", ")", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", ",", "'notifications'", ":", "unread_ntf", ",", "'messages'", ":", "unread_msg", "}" ]
Number of unread messages for current user .. code-block:: python # request: { 'view':'_zops_unread_count', } # response: { 'status': 'OK', 'code': 200, 'notifications': int, 'messages': int, }
[ "Number", "of", "unread", "messages", "for", "current", "user" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L305-L342
zetaops/zengine
zengine/messaging/views.py
get_notifications
def get_notifications(current): """ Returns last N notifications for current user .. code-block:: python # request: { 'view':'_zops_unread_messages', 'amount': int, # Optional, defaults to 8 } # response: { 'status': 'OK', 'code': 200, 'notifications': [{'title':string, 'body': string, 'channel_key': key, 'type': int, 'url': string, # could be a in app JS URL prefixed with "#" or # full blown URL prefixed with "http" 'message_key': key, 'timestamp': datetime},], } """ current.output = { 'status': 'OK', 'code': 200, 'notifications': [], } amount = current.input.get('amount', 8) try: notif_sbs = current.user.subscriptions.objects.get(channel_id=current.user.prv_exchange) except MultipleObjectsReturned: # FIXME: This should not happen, log.exception("MULTIPLE PRV EXCHANGES!!!!") sbs = current.user.subscriptions.objects.filter(channel_id=current.user.prv_exchange) sbs[0].delete() notif_sbs = sbs[1] for msg in notif_sbs.channel.message_set.objects.all()[:amount]: current.output['notifications'].insert(0, { 'title': msg.msg_title, 'body': msg.body, 'type': msg.typ, 'url': msg.url, 'channel_key': msg.channel.key, 'message_key': msg.key, 'timestamp': msg.updated_at})
python
def get_notifications(current): """ Returns last N notifications for current user .. code-block:: python # request: { 'view':'_zops_unread_messages', 'amount': int, # Optional, defaults to 8 } # response: { 'status': 'OK', 'code': 200, 'notifications': [{'title':string, 'body': string, 'channel_key': key, 'type': int, 'url': string, # could be a in app JS URL prefixed with "#" or # full blown URL prefixed with "http" 'message_key': key, 'timestamp': datetime},], } """ current.output = { 'status': 'OK', 'code': 200, 'notifications': [], } amount = current.input.get('amount', 8) try: notif_sbs = current.user.subscriptions.objects.get(channel_id=current.user.prv_exchange) except MultipleObjectsReturned: # FIXME: This should not happen, log.exception("MULTIPLE PRV EXCHANGES!!!!") sbs = current.user.subscriptions.objects.filter(channel_id=current.user.prv_exchange) sbs[0].delete() notif_sbs = sbs[1] for msg in notif_sbs.channel.message_set.objects.all()[:amount]: current.output['notifications'].insert(0, { 'title': msg.msg_title, 'body': msg.body, 'type': msg.typ, 'url': msg.url, 'channel_key': msg.channel.key, 'message_key': msg.key, 'timestamp': msg.updated_at})
[ "def", "get_notifications", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", ",", "'notifications'", ":", "[", "]", ",", "}", "amount", "=", "current", ".", "input", ".", "get", "(", "'amount'", ",", "8", ")", "try", ":", "notif_sbs", "=", "current", ".", "user", ".", "subscriptions", ".", "objects", ".", "get", "(", "channel_id", "=", "current", ".", "user", ".", "prv_exchange", ")", "except", "MultipleObjectsReturned", ":", "# FIXME: This should not happen,", "log", ".", "exception", "(", "\"MULTIPLE PRV EXCHANGES!!!!\"", ")", "sbs", "=", "current", ".", "user", ".", "subscriptions", ".", "objects", ".", "filter", "(", "channel_id", "=", "current", ".", "user", ".", "prv_exchange", ")", "sbs", "[", "0", "]", ".", "delete", "(", ")", "notif_sbs", "=", "sbs", "[", "1", "]", "for", "msg", "in", "notif_sbs", ".", "channel", ".", "message_set", ".", "objects", ".", "all", "(", ")", "[", ":", "amount", "]", ":", "current", ".", "output", "[", "'notifications'", "]", ".", "insert", "(", "0", ",", "{", "'title'", ":", "msg", ".", "msg_title", ",", "'body'", ":", "msg", ".", "body", ",", "'type'", ":", "msg", ".", "typ", ",", "'url'", ":", "msg", ".", "url", ",", "'channel_key'", ":", "msg", ".", "channel", ".", "key", ",", "'message_key'", ":", "msg", ".", "key", ",", "'timestamp'", ":", "msg", ".", "updated_at", "}", ")" ]
Returns last N notifications for current user .. code-block:: python # request: { 'view':'_zops_unread_messages', 'amount': int, # Optional, defaults to 8 } # response: { 'status': 'OK', 'code': 200, 'notifications': [{'title':string, 'body': string, 'channel_key': key, 'type': int, 'url': string, # could be a in app JS URL prefixed with "#" or # full blown URL prefixed with "http" 'message_key': key, 'timestamp': datetime},], }
[ "Returns", "last", "N", "notifications", "for", "current", "user" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L345-L394
zetaops/zengine
zengine/messaging/views.py
create_channel
def create_channel(current): """ Create a public channel. Can be a broadcast channel or normal chat room. Chat room and broadcast distinction will be made at user subscription phase. .. code-block:: python # request: { 'view':'_zops_create_channel', 'name': string, 'description': string, } # response: { 'description': string, 'name': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'last_messages': [MSG_DICT] 'status': 'Created', 'code': 201, 'key': key, # of just created channel } """ channel = Channel(name=current.input['name'], description=current.input['description'], owner=current.user, typ=15).save() with BlockSave(Subscriber): Subscriber.objects.get_or_create(user=channel.owner, channel=channel, can_manage=True, can_leave=False) current.input['key'] = channel.key show_channel(current) current.output.update({ 'status': 'Created', 'code': 201 })
python
def create_channel(current): """ Create a public channel. Can be a broadcast channel or normal chat room. Chat room and broadcast distinction will be made at user subscription phase. .. code-block:: python # request: { 'view':'_zops_create_channel', 'name': string, 'description': string, } # response: { 'description': string, 'name': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'last_messages': [MSG_DICT] 'status': 'Created', 'code': 201, 'key': key, # of just created channel } """ channel = Channel(name=current.input['name'], description=current.input['description'], owner=current.user, typ=15).save() with BlockSave(Subscriber): Subscriber.objects.get_or_create(user=channel.owner, channel=channel, can_manage=True, can_leave=False) current.input['key'] = channel.key show_channel(current) current.output.update({ 'status': 'Created', 'code': 201 })
[ "def", "create_channel", "(", "current", ")", ":", "channel", "=", "Channel", "(", "name", "=", "current", ".", "input", "[", "'name'", "]", ",", "description", "=", "current", ".", "input", "[", "'description'", "]", ",", "owner", "=", "current", ".", "user", ",", "typ", "=", "15", ")", ".", "save", "(", ")", "with", "BlockSave", "(", "Subscriber", ")", ":", "Subscriber", ".", "objects", ".", "get_or_create", "(", "user", "=", "channel", ".", "owner", ",", "channel", "=", "channel", ",", "can_manage", "=", "True", ",", "can_leave", "=", "False", ")", "current", ".", "input", "[", "'key'", "]", "=", "channel", ".", "key", "show_channel", "(", "current", ")", "current", ".", "output", ".", "update", "(", "{", "'status'", ":", "'Created'", ",", "'code'", ":", "201", "}", ")" ]
Create a public channel. Can be a broadcast channel or normal chat room. Chat room and broadcast distinction will be made at user subscription phase. .. code-block:: python # request: { 'view':'_zops_create_channel', 'name': string, 'description': string, } # response: { 'description': string, 'name': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'last_messages': [MSG_DICT] 'status': 'Created', 'code': 201, 'key': key, # of just created channel }
[ "Create", "a", "public", "channel", ".", "Can", "be", "a", "broadcast", "channel", "or", "normal", "chat", "room", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L397-L442
zetaops/zengine
zengine/messaging/views.py
add_members
def add_members(current): """ Subscribe member(s) to a channel .. code-block:: python # request: { 'view':'_zops_add_members', 'channel_key': key, 'read_only': boolean, # true if this is a Broadcast channel, # false if it's a normal chat room 'members': [key, key], } # response: { 'existing': [key,], # existing members 'newly_added': [key,], # newly added members 'status': 'Created', 'code': 201 } """ newly_added, existing = [], [] read_only = current.input['read_only'] for member_key in current.input['members']: sb, new = Subscriber(current).objects.get_or_create(user_id=member_key, read_only=read_only, channel_id=current.input['channel_key']) if new: newly_added.append(member_key) else: existing.append(member_key) current.output = { 'existing': existing, 'newly_added': newly_added, 'status': 'OK', 'code': 201 }
python
def add_members(current): """ Subscribe member(s) to a channel .. code-block:: python # request: { 'view':'_zops_add_members', 'channel_key': key, 'read_only': boolean, # true if this is a Broadcast channel, # false if it's a normal chat room 'members': [key, key], } # response: { 'existing': [key,], # existing members 'newly_added': [key,], # newly added members 'status': 'Created', 'code': 201 } """ newly_added, existing = [], [] read_only = current.input['read_only'] for member_key in current.input['members']: sb, new = Subscriber(current).objects.get_or_create(user_id=member_key, read_only=read_only, channel_id=current.input['channel_key']) if new: newly_added.append(member_key) else: existing.append(member_key) current.output = { 'existing': existing, 'newly_added': newly_added, 'status': 'OK', 'code': 201 }
[ "def", "add_members", "(", "current", ")", ":", "newly_added", ",", "existing", "=", "[", "]", ",", "[", "]", "read_only", "=", "current", ".", "input", "[", "'read_only'", "]", "for", "member_key", "in", "current", ".", "input", "[", "'members'", "]", ":", "sb", ",", "new", "=", "Subscriber", "(", "current", ")", ".", "objects", ".", "get_or_create", "(", "user_id", "=", "member_key", ",", "read_only", "=", "read_only", ",", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ")", "if", "new", ":", "newly_added", ".", "append", "(", "member_key", ")", "else", ":", "existing", ".", "append", "(", "member_key", ")", "current", ".", "output", "=", "{", "'existing'", ":", "existing", ",", "'newly_added'", ":", "newly_added", ",", "'status'", ":", "'OK'", ",", "'code'", ":", "201", "}" ]
Subscribe member(s) to a channel .. code-block:: python # request: { 'view':'_zops_add_members', 'channel_key': key, 'read_only': boolean, # true if this is a Broadcast channel, # false if it's a normal chat room 'members': [key, key], } # response: { 'existing': [key,], # existing members 'newly_added': [key,], # newly added members 'status': 'Created', 'code': 201 }
[ "Subscribe", "member", "(", "s", ")", "to", "a", "channel" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L445-L484
zetaops/zengine
zengine/messaging/views.py
add_unit_to_channel
def add_unit_to_channel(current): """ Subscribe users of a given unit to given channel JSON API: .. code-block:: python # request: { 'view':'_zops_add_unit_to_channel', 'unit_key': key, 'channel_key': key, 'read_only': boolean, # true if this is a Broadcast channel, # false if it's a normal chat room } # response: { 'existing': [key,], # existing members 'newly_added': [key,], # newly added members 'status': 'Created', 'code': 201 } """ read_only = current.input['read_only'] newly_added, existing = [], [] for member_key in UnitModel.get_user_keys(current, current.input['unit_key']): sb, new = Subscriber(current).objects.get_or_create(user_id=member_key, read_only=read_only, channel_id=current.input['channel_key']) if new: newly_added.append(member_key) else: existing.append(member_key) current.output = { 'existing': existing, 'newly_added': newly_added, 'status': 'OK', 'code': 201 }
python
def add_unit_to_channel(current): """ Subscribe users of a given unit to given channel JSON API: .. code-block:: python # request: { 'view':'_zops_add_unit_to_channel', 'unit_key': key, 'channel_key': key, 'read_only': boolean, # true if this is a Broadcast channel, # false if it's a normal chat room } # response: { 'existing': [key,], # existing members 'newly_added': [key,], # newly added members 'status': 'Created', 'code': 201 } """ read_only = current.input['read_only'] newly_added, existing = [], [] for member_key in UnitModel.get_user_keys(current, current.input['unit_key']): sb, new = Subscriber(current).objects.get_or_create(user_id=member_key, read_only=read_only, channel_id=current.input['channel_key']) if new: newly_added.append(member_key) else: existing.append(member_key) current.output = { 'existing': existing, 'newly_added': newly_added, 'status': 'OK', 'code': 201 }
[ "def", "add_unit_to_channel", "(", "current", ")", ":", "read_only", "=", "current", ".", "input", "[", "'read_only'", "]", "newly_added", ",", "existing", "=", "[", "]", ",", "[", "]", "for", "member_key", "in", "UnitModel", ".", "get_user_keys", "(", "current", ",", "current", ".", "input", "[", "'unit_key'", "]", ")", ":", "sb", ",", "new", "=", "Subscriber", "(", "current", ")", ".", "objects", ".", "get_or_create", "(", "user_id", "=", "member_key", ",", "read_only", "=", "read_only", ",", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ")", "if", "new", ":", "newly_added", ".", "append", "(", "member_key", ")", "else", ":", "existing", ".", "append", "(", "member_key", ")", "current", ".", "output", "=", "{", "'existing'", ":", "existing", ",", "'newly_added'", ":", "newly_added", ",", "'status'", ":", "'OK'", ",", "'code'", ":", "201", "}" ]
Subscribe users of a given unit to given channel JSON API: .. code-block:: python # request: { 'view':'_zops_add_unit_to_channel', 'unit_key': key, 'channel_key': key, 'read_only': boolean, # true if this is a Broadcast channel, # false if it's a normal chat room } # response: { 'existing': [key,], # existing members 'newly_added': [key,], # newly added members 'status': 'Created', 'code': 201 }
[ "Subscribe", "users", "of", "a", "given", "unit", "to", "given", "channel" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L487-L528
zetaops/zengine
zengine/messaging/views.py
search_user
def search_user(current): """ Search users for adding to a public room or creating one to one direct messaging .. code-block:: python # request: { 'view':'_zops_search_user', 'query': string, } # response: { 'results': [('full_name', 'key', 'avatar_url'), ], 'status': 'OK', 'code': 200 } """ current.output = { 'results': [], 'status': 'OK', 'code': 201 } qs = UserModel(current).objects.exclude(key=current.user_id).search_on( *settings.MESSAGING_USER_SEARCH_FIELDS, contains=current.input['query']) # FIXME: somehow exclude(key=current.user_id) not working with search_on() for user in qs: if user.key != current.user_id: current.output['results'].append((user.full_name, user.key, user.get_avatar_url()))
python
def search_user(current): """ Search users for adding to a public room or creating one to one direct messaging .. code-block:: python # request: { 'view':'_zops_search_user', 'query': string, } # response: { 'results': [('full_name', 'key', 'avatar_url'), ], 'status': 'OK', 'code': 200 } """ current.output = { 'results': [], 'status': 'OK', 'code': 201 } qs = UserModel(current).objects.exclude(key=current.user_id).search_on( *settings.MESSAGING_USER_SEARCH_FIELDS, contains=current.input['query']) # FIXME: somehow exclude(key=current.user_id) not working with search_on() for user in qs: if user.key != current.user_id: current.output['results'].append((user.full_name, user.key, user.get_avatar_url()))
[ "def", "search_user", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'results'", ":", "[", "]", ",", "'status'", ":", "'OK'", ",", "'code'", ":", "201", "}", "qs", "=", "UserModel", "(", "current", ")", ".", "objects", ".", "exclude", "(", "key", "=", "current", ".", "user_id", ")", ".", "search_on", "(", "*", "settings", ".", "MESSAGING_USER_SEARCH_FIELDS", ",", "contains", "=", "current", ".", "input", "[", "'query'", "]", ")", "# FIXME: somehow exclude(key=current.user_id) not working with search_on()", "for", "user", "in", "qs", ":", "if", "user", ".", "key", "!=", "current", ".", "user_id", ":", "current", ".", "output", "[", "'results'", "]", ".", "append", "(", "(", "user", ".", "full_name", ",", "user", ".", "key", ",", "user", ".", "get_avatar_url", "(", ")", ")", ")" ]
Search users for adding to a public room or creating one to one direct messaging .. code-block:: python # request: { 'view':'_zops_search_user', 'query': string, } # response: { 'results': [('full_name', 'key', 'avatar_url'), ], 'status': 'OK', 'code': 200 }
[ "Search", "users", "for", "adding", "to", "a", "public", "room", "or", "creating", "one", "to", "one", "direct", "messaging" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L531-L563
zetaops/zengine
zengine/messaging/views.py
search_unit
def search_unit(current): """ Search on units for subscribing it's users to a channel .. code-block:: python # request: { 'view':'_zops_search_unit', 'query': string, } # response: { 'results': [('name', 'key'), ], 'status': 'OK', 'code': 200 } """ current.output = { 'results': [], 'status': 'OK', 'code': 201 } for user in UnitModel(current).objects.search_on(*settings.MESSAGING_UNIT_SEARCH_FIELDS, contains=current.input['query']): current.output['results'].append((user.name, user.key))
python
def search_unit(current): """ Search on units for subscribing it's users to a channel .. code-block:: python # request: { 'view':'_zops_search_unit', 'query': string, } # response: { 'results': [('name', 'key'), ], 'status': 'OK', 'code': 200 } """ current.output = { 'results': [], 'status': 'OK', 'code': 201 } for user in UnitModel(current).objects.search_on(*settings.MESSAGING_UNIT_SEARCH_FIELDS, contains=current.input['query']): current.output['results'].append((user.name, user.key))
[ "def", "search_unit", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'results'", ":", "[", "]", ",", "'status'", ":", "'OK'", ",", "'code'", ":", "201", "}", "for", "user", "in", "UnitModel", "(", "current", ")", ".", "objects", ".", "search_on", "(", "*", "settings", ".", "MESSAGING_UNIT_SEARCH_FIELDS", ",", "contains", "=", "current", ".", "input", "[", "'query'", "]", ")", ":", "current", ".", "output", "[", "'results'", "]", ".", "append", "(", "(", "user", ".", "name", ",", "user", ".", "key", ")", ")" ]
Search on units for subscribing it's users to a channel .. code-block:: python # request: { 'view':'_zops_search_unit', 'query': string, } # response: { 'results': [('name', 'key'), ], 'status': 'OK', 'code': 200 }
[ "Search", "on", "units", "for", "subscribing", "it", "s", "users", "to", "a", "channel" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L566-L592
zetaops/zengine
zengine/messaging/views.py
create_direct_channel
def create_direct_channel(current): """ Create a One-To-One channel between current and selected user. .. code-block:: python # request: { 'view':'_zops_create_direct_channel', 'user_key': key, } # response: { 'description': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'last_messages': [MSG_DICT] 'status': 'Created', 'code': 201, 'channel_key': key, # of just created channel 'name': string, # name of subscribed channel } """ channel, sub_name = Channel.get_or_create_direct_channel(current.user_id, current.input['user_key']) current.input['key'] = channel.key show_channel(current) current.output.update({ 'status': 'Created', 'code': 201 })
python
def create_direct_channel(current): """ Create a One-To-One channel between current and selected user. .. code-block:: python # request: { 'view':'_zops_create_direct_channel', 'user_key': key, } # response: { 'description': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'last_messages': [MSG_DICT] 'status': 'Created', 'code': 201, 'channel_key': key, # of just created channel 'name': string, # name of subscribed channel } """ channel, sub_name = Channel.get_or_create_direct_channel(current.user_id, current.input['user_key']) current.input['key'] = channel.key show_channel(current) current.output.update({ 'status': 'Created', 'code': 201 })
[ "def", "create_direct_channel", "(", "current", ")", ":", "channel", ",", "sub_name", "=", "Channel", ".", "get_or_create_direct_channel", "(", "current", ".", "user_id", ",", "current", ".", "input", "[", "'user_key'", "]", ")", "current", ".", "input", "[", "'key'", "]", "=", "channel", ".", "key", "show_channel", "(", "current", ")", "current", ".", "output", ".", "update", "(", "{", "'status'", ":", "'Created'", ",", "'code'", ":", "201", "}", ")" ]
Create a One-To-One channel between current and selected user. .. code-block:: python # request: { 'view':'_zops_create_direct_channel', 'user_key': key, } # response: { 'description': string, 'no_of_members': int, 'member_list': [ {'name': string, 'is_online': bool, 'avatar_url': string, }], 'last_messages': [MSG_DICT] 'status': 'Created', 'code': 201, 'channel_key': key, # of just created channel 'name': string, # name of subscribed channel }
[ "Create", "a", "One", "-", "To", "-", "One", "channel", "between", "current", "and", "selected", "user", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L595-L631
zetaops/zengine
zengine/messaging/views.py
find_message
def find_message(current): """ Search in messages. If "channel_key" given, search will be limited to that channel, otherwise search will be performed on all of user's subscribed channels. .. code-block:: python # request: { 'view':'_zops_search_unit, 'channel_key': key, 'query': string, 'page': int, } # response: { 'results': [MSG_DICT, ], 'pagination': { 'page': int, # current page 'total_pages': int, 'total_objects': int, 'per_page': int, # object per page }, 'status': 'OK', 'code': 200 } """ current.output = { 'results': [], 'status': 'OK', 'code': 201 } query_set = Message(current).objects.search_on(['msg_title', 'body', 'url'], contains=current.input['query']) if current.input['channel_key']: query_set = query_set.filter(channel_id=current.input['channel_key']) else: subscribed_channels = Subscriber.objects.filter(user_id=current.user_id).values_list( "channel_id", flatten=True) query_set = query_set.filter(channel_id__in=subscribed_channels) query_set, pagination_data = _paginate(current_page=current.input['page'], query_set=query_set) current.output['pagination'] = pagination_data for msg in query_set: current.output['results'].append(msg.serialize(current.user))
python
def find_message(current): """ Search in messages. If "channel_key" given, search will be limited to that channel, otherwise search will be performed on all of user's subscribed channels. .. code-block:: python # request: { 'view':'_zops_search_unit, 'channel_key': key, 'query': string, 'page': int, } # response: { 'results': [MSG_DICT, ], 'pagination': { 'page': int, # current page 'total_pages': int, 'total_objects': int, 'per_page': int, # object per page }, 'status': 'OK', 'code': 200 } """ current.output = { 'results': [], 'status': 'OK', 'code': 201 } query_set = Message(current).objects.search_on(['msg_title', 'body', 'url'], contains=current.input['query']) if current.input['channel_key']: query_set = query_set.filter(channel_id=current.input['channel_key']) else: subscribed_channels = Subscriber.objects.filter(user_id=current.user_id).values_list( "channel_id", flatten=True) query_set = query_set.filter(channel_id__in=subscribed_channels) query_set, pagination_data = _paginate(current_page=current.input['page'], query_set=query_set) current.output['pagination'] = pagination_data for msg in query_set: current.output['results'].append(msg.serialize(current.user))
[ "def", "find_message", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'results'", ":", "[", "]", ",", "'status'", ":", "'OK'", ",", "'code'", ":", "201", "}", "query_set", "=", "Message", "(", "current", ")", ".", "objects", ".", "search_on", "(", "[", "'msg_title'", ",", "'body'", ",", "'url'", "]", ",", "contains", "=", "current", ".", "input", "[", "'query'", "]", ")", "if", "current", ".", "input", "[", "'channel_key'", "]", ":", "query_set", "=", "query_set", ".", "filter", "(", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ")", "else", ":", "subscribed_channels", "=", "Subscriber", ".", "objects", ".", "filter", "(", "user_id", "=", "current", ".", "user_id", ")", ".", "values_list", "(", "\"channel_id\"", ",", "flatten", "=", "True", ")", "query_set", "=", "query_set", ".", "filter", "(", "channel_id__in", "=", "subscribed_channels", ")", "query_set", ",", "pagination_data", "=", "_paginate", "(", "current_page", "=", "current", ".", "input", "[", "'page'", "]", ",", "query_set", "=", "query_set", ")", "current", ".", "output", "[", "'pagination'", "]", "=", "pagination_data", "for", "msg", "in", "query_set", ":", "current", ".", "output", "[", "'results'", "]", ".", "append", "(", "msg", ".", "serialize", "(", "current", ".", "user", ")", ")" ]
Search in messages. If "channel_key" given, search will be limited to that channel, otherwise search will be performed on all of user's subscribed channels. .. code-block:: python # request: { 'view':'_zops_search_unit, 'channel_key': key, 'query': string, 'page': int, } # response: { 'results': [MSG_DICT, ], 'pagination': { 'page': int, # current page 'total_pages': int, 'total_objects': int, 'per_page': int, # object per page }, 'status': 'OK', 'code': 200 }
[ "Search", "in", "messages", ".", "If", "channel_key", "given", "search", "will", "be", "limited", "to", "that", "channel", "otherwise", "search", "will", "be", "performed", "on", "all", "of", "user", "s", "subscribed", "channels", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L634-L679
zetaops/zengine
zengine/messaging/views.py
delete_channel
def delete_channel(current): """ Delete a channel .. code-block:: python # request: { 'view':'_zops_delete_channel, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 } """ ch_key = current.input['channel_key'] ch = Channel(current).objects.get(owner_id=current.user_id, key=ch_key) ch.delete() Subscriber.objects.filter(channel_id=ch_key).delete() Message.objects.filter(channel_id=ch_key).delete() current.output = {'status': 'Deleted', 'code': 200}
python
def delete_channel(current): """ Delete a channel .. code-block:: python # request: { 'view':'_zops_delete_channel, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 } """ ch_key = current.input['channel_key'] ch = Channel(current).objects.get(owner_id=current.user_id, key=ch_key) ch.delete() Subscriber.objects.filter(channel_id=ch_key).delete() Message.objects.filter(channel_id=ch_key).delete() current.output = {'status': 'Deleted', 'code': 200}
[ "def", "delete_channel", "(", "current", ")", ":", "ch_key", "=", "current", ".", "input", "[", "'channel_key'", "]", "ch", "=", "Channel", "(", "current", ")", ".", "objects", ".", "get", "(", "owner_id", "=", "current", ".", "user_id", ",", "key", "=", "ch_key", ")", "ch", ".", "delete", "(", ")", "Subscriber", ".", "objects", ".", "filter", "(", "channel_id", "=", "ch_key", ")", ".", "delete", "(", ")", "Message", ".", "objects", ".", "filter", "(", "channel_id", "=", "ch_key", ")", ".", "delete", "(", ")", "current", ".", "output", "=", "{", "'status'", ":", "'Deleted'", ",", "'code'", ":", "200", "}" ]
Delete a channel .. code-block:: python # request: { 'view':'_zops_delete_channel, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 }
[ "Delete", "a", "channel" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L682-L706
zetaops/zengine
zengine/messaging/views.py
edit_channel
def edit_channel(current): """ Update channel name or description .. code-block:: python # request: { 'view':'_zops_edit_channel, 'channel_key': key, 'name': string, 'description': string, } # response: { 'status': 'OK', 'code': 200 } """ ch = Channel(current).objects.get(owner_id=current.user_id, key=current.input['channel_key']) ch.name = current.input['name'] ch.description = current.input['description'] ch.save() for sbs in ch.subscriber_set.objects.all(): sbs.name = ch.name sbs.save() current.output = {'status': 'OK', 'code': 200}
python
def edit_channel(current): """ Update channel name or description .. code-block:: python # request: { 'view':'_zops_edit_channel, 'channel_key': key, 'name': string, 'description': string, } # response: { 'status': 'OK', 'code': 200 } """ ch = Channel(current).objects.get(owner_id=current.user_id, key=current.input['channel_key']) ch.name = current.input['name'] ch.description = current.input['description'] ch.save() for sbs in ch.subscriber_set.objects.all(): sbs.name = ch.name sbs.save() current.output = {'status': 'OK', 'code': 200}
[ "def", "edit_channel", "(", "current", ")", ":", "ch", "=", "Channel", "(", "current", ")", ".", "objects", ".", "get", "(", "owner_id", "=", "current", ".", "user_id", ",", "key", "=", "current", ".", "input", "[", "'channel_key'", "]", ")", "ch", ".", "name", "=", "current", ".", "input", "[", "'name'", "]", "ch", ".", "description", "=", "current", ".", "input", "[", "'description'", "]", "ch", ".", "save", "(", ")", "for", "sbs", "in", "ch", ".", "subscriber_set", ".", "objects", ".", "all", "(", ")", ":", "sbs", ".", "name", "=", "ch", ".", "name", "sbs", ".", "save", "(", ")", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}" ]
Update channel name or description .. code-block:: python # request: { 'view':'_zops_edit_channel, 'channel_key': key, 'name': string, 'description': string, } # response: { 'status': 'OK', 'code': 200 }
[ "Update", "channel", "name", "or", "description" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L709-L737
zetaops/zengine
zengine/messaging/views.py
pin_channel
def pin_channel(current): """ Pin a channel to top of channel list .. code-block:: python # request: { 'view':'_zops_pin_channel, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 } """ try: Subscriber(current).objects.filter(user_id=current.user_id, channel_id=current.input['channel_key']).update( pinned=True) current.output = {'status': 'OK', 'code': 200} except ObjectDoesNotExist: raise HTTPError(404, "")
python
def pin_channel(current): """ Pin a channel to top of channel list .. code-block:: python # request: { 'view':'_zops_pin_channel, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 } """ try: Subscriber(current).objects.filter(user_id=current.user_id, channel_id=current.input['channel_key']).update( pinned=True) current.output = {'status': 'OK', 'code': 200} except ObjectDoesNotExist: raise HTTPError(404, "")
[ "def", "pin_channel", "(", "current", ")", ":", "try", ":", "Subscriber", "(", "current", ")", ".", "objects", ".", "filter", "(", "user_id", "=", "current", ".", "user_id", ",", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ")", ".", "update", "(", "pinned", "=", "True", ")", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}", "except", "ObjectDoesNotExist", ":", "raise", "HTTPError", "(", "404", ",", "\"\"", ")" ]
Pin a channel to top of channel list .. code-block:: python # request: { 'view':'_zops_pin_channel, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 }
[ "Pin", "a", "channel", "to", "top", "of", "channel", "list" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L740-L764
zetaops/zengine
zengine/messaging/views.py
delete_message
def delete_message(current): """ Delete a message .. code-block:: python # request: { 'view':'_zops_delete_message, 'message_key': key, } # response: { 'key': key, 'status': 'OK', 'code': 200 } """ try: Message(current).objects.get(sender_id=current.user_id, key=current.input['key']).delete() current.output = {'status': 'Deleted', 'code': 200, 'key': current.input['key']} except ObjectDoesNotExist: raise HTTPError(404, "")
python
def delete_message(current): """ Delete a message .. code-block:: python # request: { 'view':'_zops_delete_message, 'message_key': key, } # response: { 'key': key, 'status': 'OK', 'code': 200 } """ try: Message(current).objects.get(sender_id=current.user_id, key=current.input['key']).delete() current.output = {'status': 'Deleted', 'code': 200, 'key': current.input['key']} except ObjectDoesNotExist: raise HTTPError(404, "")
[ "def", "delete_message", "(", "current", ")", ":", "try", ":", "Message", "(", "current", ")", ".", "objects", ".", "get", "(", "sender_id", "=", "current", ".", "user_id", ",", "key", "=", "current", ".", "input", "[", "'key'", "]", ")", ".", "delete", "(", ")", "current", ".", "output", "=", "{", "'status'", ":", "'Deleted'", ",", "'code'", ":", "200", ",", "'key'", ":", "current", ".", "input", "[", "'key'", "]", "}", "except", "ObjectDoesNotExist", ":", "raise", "HTTPError", "(", "404", ",", "\"\"", ")" ]
Delete a message .. code-block:: python # request: { 'view':'_zops_delete_message, 'message_key': key, } # response: { 'key': key, 'status': 'OK', 'code': 200 }
[ "Delete", "a", "message" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L767-L791
zetaops/zengine
zengine/messaging/views.py
edit_message
def edit_message(current): """ Edit a message a user own. .. code-block:: python # request: { 'view':'_zops_edit_message', 'message': { 'body': string, # message text 'key': key } } # response: { 'status': string, # 'OK' for success 'code': int, # 200 for success } """ current.output = {'status': 'OK', 'code': 200} in_msg = current.input['message'] try: msg = Message(current).objects.get(sender_id=current.user_id, key=in_msg['key']) msg.body = in_msg['body'] msg.save() except ObjectDoesNotExist: raise HTTPError(404, "")
python
def edit_message(current): """ Edit a message a user own. .. code-block:: python # request: { 'view':'_zops_edit_message', 'message': { 'body': string, # message text 'key': key } } # response: { 'status': string, # 'OK' for success 'code': int, # 200 for success } """ current.output = {'status': 'OK', 'code': 200} in_msg = current.input['message'] try: msg = Message(current).objects.get(sender_id=current.user_id, key=in_msg['key']) msg.body = in_msg['body'] msg.save() except ObjectDoesNotExist: raise HTTPError(404, "")
[ "def", "edit_message", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}", "in_msg", "=", "current", ".", "input", "[", "'message'", "]", "try", ":", "msg", "=", "Message", "(", "current", ")", ".", "objects", ".", "get", "(", "sender_id", "=", "current", ".", "user_id", ",", "key", "=", "in_msg", "[", "'key'", "]", ")", "msg", ".", "body", "=", "in_msg", "[", "'body'", "]", "msg", ".", "save", "(", ")", "except", "ObjectDoesNotExist", ":", "raise", "HTTPError", "(", "404", ",", "\"\"", ")" ]
Edit a message a user own. .. code-block:: python # request: { 'view':'_zops_edit_message', 'message': { 'body': string, # message text 'key': key } } # response: { 'status': string, # 'OK' for success 'code': int, # 200 for success }
[ "Edit", "a", "message", "a", "user", "own", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L794-L822
zetaops/zengine
zengine/messaging/views.py
flag_message
def flag_message(current): """ Flag inappropriate messages .. code-block:: python # request: { 'view':'_zops_flag_message', 'message_key': key, } # response: { ' 'status': 'Created', 'code': 201, } """ current.output = {'status': 'Created', 'code': 201} FlaggedMessage.objects.get_or_create(user_id=current.user_id, message_id=current.input['key'])
python
def flag_message(current): """ Flag inappropriate messages .. code-block:: python # request: { 'view':'_zops_flag_message', 'message_key': key, } # response: { ' 'status': 'Created', 'code': 201, } """ current.output = {'status': 'Created', 'code': 201} FlaggedMessage.objects.get_or_create(user_id=current.user_id, message_id=current.input['key'])
[ "def", "flag_message", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'Created'", ",", "'code'", ":", "201", "}", "FlaggedMessage", ".", "objects", ".", "get_or_create", "(", "user_id", "=", "current", ".", "user_id", ",", "message_id", "=", "current", ".", "input", "[", "'key'", "]", ")" ]
Flag inappropriate messages .. code-block:: python # request: { 'view':'_zops_flag_message', 'message_key': key, } # response: { ' 'status': 'Created', 'code': 201, }
[ "Flag", "inappropriate", "messages" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L825-L846
zetaops/zengine
zengine/messaging/views.py
unflag_message
def unflag_message(current): """ remove flag of a message .. code-block:: python # request: { 'view':'_zops_flag_message', 'key': key, } # response: { ' 'status': 'OK', 'code': 200, } """ current.output = {'status': 'OK', 'code': 200} FlaggedMessage(current).objects.filter(user_id=current.user_id, message_id=current.input['key']).delete()
python
def unflag_message(current): """ remove flag of a message .. code-block:: python # request: { 'view':'_zops_flag_message', 'key': key, } # response: { ' 'status': 'OK', 'code': 200, } """ current.output = {'status': 'OK', 'code': 200} FlaggedMessage(current).objects.filter(user_id=current.user_id, message_id=current.input['key']).delete()
[ "def", "unflag_message", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}", "FlaggedMessage", "(", "current", ")", ".", "objects", ".", "filter", "(", "user_id", "=", "current", ".", "user_id", ",", "message_id", "=", "current", ".", "input", "[", "'key'", "]", ")", ".", "delete", "(", ")" ]
remove flag of a message .. code-block:: python # request: { 'view':'_zops_flag_message', 'key': key, } # response: { ' 'status': 'OK', 'code': 200, }
[ "remove", "flag", "of", "a", "message" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L849-L871
zetaops/zengine
zengine/messaging/views.py
get_message_actions
def get_message_actions(current): """ Returns applicable actions for current user for given message key .. code-block:: python # request: { 'view':'_zops_get_message_actions', 'key': key, } # response: { 'actions':[('name_string', 'cmd_string'),] 'status': string, # 'OK' for success 'code': int, # 200 for success } """ current.output = {'status': 'OK', 'code': 200, 'actions': Message.objects.get( current.input['key']).get_actions_for(current.user)}
python
def get_message_actions(current): """ Returns applicable actions for current user for given message key .. code-block:: python # request: { 'view':'_zops_get_message_actions', 'key': key, } # response: { 'actions':[('name_string', 'cmd_string'),] 'status': string, # 'OK' for success 'code': int, # 200 for success } """ current.output = {'status': 'OK', 'code': 200, 'actions': Message.objects.get( current.input['key']).get_actions_for(current.user)}
[ "def", "get_message_actions", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", ",", "'actions'", ":", "Message", ".", "objects", ".", "get", "(", "current", ".", "input", "[", "'key'", "]", ")", ".", "get_actions_for", "(", "current", ".", "user", ")", "}" ]
Returns applicable actions for current user for given message key .. code-block:: python # request: { 'view':'_zops_get_message_actions', 'key': key, } # response: { 'actions':[('name_string', 'cmd_string'),] 'status': string, # 'OK' for success 'code': int, # 200 for success }
[ "Returns", "applicable", "actions", "for", "current", "user", "for", "given", "message", "key" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L874-L896
zetaops/zengine
zengine/messaging/views.py
add_to_favorites
def add_to_favorites(current): """ Favorite a message .. code-block:: python # request: { 'view':'_zops_add_to_favorites, 'key': key, } # response: { 'status': 'Created', 'code': 201 'favorite_key': key } """ msg = Message.objects.get(current.input['key']) current.output = {'status': 'Created', 'code': 201} fav, new = Favorite.objects.get_or_create(user_id=current.user_id, message=msg) current.output['favorite_key'] = fav.key
python
def add_to_favorites(current): """ Favorite a message .. code-block:: python # request: { 'view':'_zops_add_to_favorites, 'key': key, } # response: { 'status': 'Created', 'code': 201 'favorite_key': key } """ msg = Message.objects.get(current.input['key']) current.output = {'status': 'Created', 'code': 201} fav, new = Favorite.objects.get_or_create(user_id=current.user_id, message=msg) current.output['favorite_key'] = fav.key
[ "def", "add_to_favorites", "(", "current", ")", ":", "msg", "=", "Message", ".", "objects", ".", "get", "(", "current", ".", "input", "[", "'key'", "]", ")", "current", ".", "output", "=", "{", "'status'", ":", "'Created'", ",", "'code'", ":", "201", "}", "fav", ",", "new", "=", "Favorite", ".", "objects", ".", "get_or_create", "(", "user_id", "=", "current", ".", "user_id", ",", "message", "=", "msg", ")", "current", ".", "output", "[", "'favorite_key'", "]", "=", "fav", ".", "key" ]
Favorite a message .. code-block:: python # request: { 'view':'_zops_add_to_favorites, 'key': key, } # response: { 'status': 'Created', 'code': 201 'favorite_key': key }
[ "Favorite", "a", "message" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L899-L922
zetaops/zengine
zengine/messaging/views.py
remove_from_favorites
def remove_from_favorites(current): """ Remove a message from favorites .. code-block:: python # request: { 'view':'_zops_remove_from_favorites, 'key': key, } # response: { 'status': 'OK', 'code': 200 } """ try: current.output = {'status': 'OK', 'code': 200} Favorite(current).objects.get(user_id=current.user_id, key=current.input['key']).delete() except ObjectDoesNotExist: raise HTTPError(404, "")
python
def remove_from_favorites(current): """ Remove a message from favorites .. code-block:: python # request: { 'view':'_zops_remove_from_favorites, 'key': key, } # response: { 'status': 'OK', 'code': 200 } """ try: current.output = {'status': 'OK', 'code': 200} Favorite(current).objects.get(user_id=current.user_id, key=current.input['key']).delete() except ObjectDoesNotExist: raise HTTPError(404, "")
[ "def", "remove_from_favorites", "(", "current", ")", ":", "try", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", "}", "Favorite", "(", "current", ")", ".", "objects", ".", "get", "(", "user_id", "=", "current", ".", "user_id", ",", "key", "=", "current", ".", "input", "[", "'key'", "]", ")", ".", "delete", "(", ")", "except", "ObjectDoesNotExist", ":", "raise", "HTTPError", "(", "404", ",", "\"\"", ")" ]
Remove a message from favorites .. code-block:: python # request: { 'view':'_zops_remove_from_favorites, 'key': key, } # response: { 'status': 'OK', 'code': 200 }
[ "Remove", "a", "message", "from", "favorites" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L925-L949
zetaops/zengine
zengine/messaging/views.py
list_favorites
def list_favorites(current): """ List user's favorites. If "channel_key" given, will return favorites belong to that channel. .. code-block:: python # request: { 'view':'_zops_list_favorites, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 'favorites':[{'key': key, 'channel_key': key, 'message_key': key, 'message_summary': string, # max 60 char 'channel_name': string, },] } """ current.output = {'status': 'OK', 'code': 200, 'favorites': []} query_set = Favorite(current).objects.filter(user_id=current.user_id) if current.input['channel_key']: query_set = query_set.filter(channel_id=current.input['channel_key']) current.output['favorites'] = [{ 'key': fav.key, 'channel_key': fav.channel.key, 'message_key': fav.message.key, 'message_summary': fav.summary, 'channel_name': fav.channel_name } for fav in query_set]
python
def list_favorites(current): """ List user's favorites. If "channel_key" given, will return favorites belong to that channel. .. code-block:: python # request: { 'view':'_zops_list_favorites, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 'favorites':[{'key': key, 'channel_key': key, 'message_key': key, 'message_summary': string, # max 60 char 'channel_name': string, },] } """ current.output = {'status': 'OK', 'code': 200, 'favorites': []} query_set = Favorite(current).objects.filter(user_id=current.user_id) if current.input['channel_key']: query_set = query_set.filter(channel_id=current.input['channel_key']) current.output['favorites'] = [{ 'key': fav.key, 'channel_key': fav.channel.key, 'message_key': fav.message.key, 'message_summary': fav.summary, 'channel_name': fav.channel_name } for fav in query_set]
[ "def", "list_favorites", "(", "current", ")", ":", "current", ".", "output", "=", "{", "'status'", ":", "'OK'", ",", "'code'", ":", "200", ",", "'favorites'", ":", "[", "]", "}", "query_set", "=", "Favorite", "(", "current", ")", ".", "objects", ".", "filter", "(", "user_id", "=", "current", ".", "user_id", ")", "if", "current", ".", "input", "[", "'channel_key'", "]", ":", "query_set", "=", "query_set", ".", "filter", "(", "channel_id", "=", "current", ".", "input", "[", "'channel_key'", "]", ")", "current", ".", "output", "[", "'favorites'", "]", "=", "[", "{", "'key'", ":", "fav", ".", "key", ",", "'channel_key'", ":", "fav", ".", "channel", ".", "key", ",", "'message_key'", ":", "fav", ".", "message", ".", "key", ",", "'message_summary'", ":", "fav", ".", "summary", ",", "'channel_name'", ":", "fav", ".", "channel_name", "}", "for", "fav", "in", "query_set", "]" ]
List user's favorites. If "channel_key" given, will return favorites belong to that channel. .. code-block:: python # request: { 'view':'_zops_list_favorites, 'channel_key': key, } # response: { 'status': 'OK', 'code': 200 'favorites':[{'key': key, 'channel_key': key, 'message_key': key, 'message_summary': string, # max 60 char 'channel_name': string, },] }
[ "List", "user", "s", "favorites", ".", "If", "channel_key", "given", "will", "return", "favorites", "belong", "to", "that", "channel", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L952-L987
zetaops/zengine
zengine/messaging/model.py
Channel.get_or_create_direct_channel
def get_or_create_direct_channel(cls, initiator_key, receiver_key): """ Creates a direct messaging channel between two user Args: initiator: User, who want's to make first contact receiver: User, other party Returns: (Channel, receiver_name) """ existing = cls.objects.OR().filter( code_name='%s_%s' % (initiator_key, receiver_key)).filter( code_name='%s_%s' % (receiver_key, initiator_key)) receiver_name = UserModel.objects.get(receiver_key).full_name if existing: channel = existing[0] else: channel_name = '%s_%s' % (initiator_key, receiver_key) channel = cls(is_direct=True, code_name=channel_name, typ=10).blocking_save() with BlockSave(Subscriber): Subscriber.objects.get_or_create(channel=channel, user_id=initiator_key, name=receiver_name) Subscriber.objects.get_or_create(channel=channel, user_id=receiver_key, name=UserModel.objects.get(initiator_key).full_name) return channel, receiver_name
python
def get_or_create_direct_channel(cls, initiator_key, receiver_key): """ Creates a direct messaging channel between two user Args: initiator: User, who want's to make first contact receiver: User, other party Returns: (Channel, receiver_name) """ existing = cls.objects.OR().filter( code_name='%s_%s' % (initiator_key, receiver_key)).filter( code_name='%s_%s' % (receiver_key, initiator_key)) receiver_name = UserModel.objects.get(receiver_key).full_name if existing: channel = existing[0] else: channel_name = '%s_%s' % (initiator_key, receiver_key) channel = cls(is_direct=True, code_name=channel_name, typ=10).blocking_save() with BlockSave(Subscriber): Subscriber.objects.get_or_create(channel=channel, user_id=initiator_key, name=receiver_name) Subscriber.objects.get_or_create(channel=channel, user_id=receiver_key, name=UserModel.objects.get(initiator_key).full_name) return channel, receiver_name
[ "def", "get_or_create_direct_channel", "(", "cls", ",", "initiator_key", ",", "receiver_key", ")", ":", "existing", "=", "cls", ".", "objects", ".", "OR", "(", ")", ".", "filter", "(", "code_name", "=", "'%s_%s'", "%", "(", "initiator_key", ",", "receiver_key", ")", ")", ".", "filter", "(", "code_name", "=", "'%s_%s'", "%", "(", "receiver_key", ",", "initiator_key", ")", ")", "receiver_name", "=", "UserModel", ".", "objects", ".", "get", "(", "receiver_key", ")", ".", "full_name", "if", "existing", ":", "channel", "=", "existing", "[", "0", "]", "else", ":", "channel_name", "=", "'%s_%s'", "%", "(", "initiator_key", ",", "receiver_key", ")", "channel", "=", "cls", "(", "is_direct", "=", "True", ",", "code_name", "=", "channel_name", ",", "typ", "=", "10", ")", ".", "blocking_save", "(", ")", "with", "BlockSave", "(", "Subscriber", ")", ":", "Subscriber", ".", "objects", ".", "get_or_create", "(", "channel", "=", "channel", ",", "user_id", "=", "initiator_key", ",", "name", "=", "receiver_name", ")", "Subscriber", ".", "objects", ".", "get_or_create", "(", "channel", "=", "channel", ",", "user_id", "=", "receiver_key", ",", "name", "=", "UserModel", ".", "objects", ".", "get", "(", "initiator_key", ")", ".", "full_name", ")", "return", "channel", ",", "receiver_name" ]
Creates a direct messaging channel between two user Args: initiator: User, who want's to make first contact receiver: User, other party Returns: (Channel, receiver_name)
[ "Creates", "a", "direct", "messaging", "channel", "between", "two", "user" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L75-L102
zetaops/zengine
zengine/messaging/model.py
Channel.create_exchange
def create_exchange(self): """ Creates MQ exchange for this channel Needs to be defined only once. """ mq_channel = self._connect_mq() mq_channel.exchange_declare(exchange=self.code_name, exchange_type='fanout', durable=True)
python
def create_exchange(self): """ Creates MQ exchange for this channel Needs to be defined only once. """ mq_channel = self._connect_mq() mq_channel.exchange_declare(exchange=self.code_name, exchange_type='fanout', durable=True)
[ "def", "create_exchange", "(", "self", ")", ":", "mq_channel", "=", "self", ".", "_connect_mq", "(", ")", "mq_channel", ".", "exchange_declare", "(", "exchange", "=", "self", ".", "code_name", ",", "exchange_type", "=", "'fanout'", ",", "durable", "=", "True", ")" ]
Creates MQ exchange for this channel Needs to be defined only once.
[ "Creates", "MQ", "exchange", "for", "this", "channel", "Needs", "to", "be", "defined", "only", "once", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L135-L143
zetaops/zengine
zengine/messaging/model.py
Channel.delete_exchange
def delete_exchange(self): """ Deletes MQ exchange for this channel Needs to be defined only once. """ mq_channel = self._connect_mq() mq_channel.exchange_delete(exchange=self.code_name)
python
def delete_exchange(self): """ Deletes MQ exchange for this channel Needs to be defined only once. """ mq_channel = self._connect_mq() mq_channel.exchange_delete(exchange=self.code_name)
[ "def", "delete_exchange", "(", "self", ")", ":", "mq_channel", "=", "self", ".", "_connect_mq", "(", ")", "mq_channel", ".", "exchange_delete", "(", "exchange", "=", "self", ".", "code_name", ")" ]
Deletes MQ exchange for this channel Needs to be defined only once.
[ "Deletes", "MQ", "exchange", "for", "this", "channel", "Needs", "to", "be", "defined", "only", "once", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L145-L151
zetaops/zengine
zengine/messaging/model.py
Subscriber.get_channel_listing
def get_channel_listing(self): """ serialized form for channel listing """ return {'name': self.name, 'key': self.channel.key, 'type': self.channel.typ, 'read_only': self.read_only, 'is_online': self.is_online(), 'actions': self.get_actions(), 'unread': self.unread_count()}
python
def get_channel_listing(self): """ serialized form for channel listing """ return {'name': self.name, 'key': self.channel.key, 'type': self.channel.typ, 'read_only': self.read_only, 'is_online': self.is_online(), 'actions': self.get_actions(), 'unread': self.unread_count()}
[ "def", "get_channel_listing", "(", "self", ")", ":", "return", "{", "'name'", ":", "self", ".", "name", ",", "'key'", ":", "self", ".", "channel", ".", "key", ",", "'type'", ":", "self", ".", "channel", ".", "typ", ",", "'read_only'", ":", "self", ".", "read_only", ",", "'is_online'", ":", "self", ".", "is_online", "(", ")", ",", "'actions'", ":", "self", ".", "get_actions", "(", ")", ",", "'unread'", ":", "self", ".", "unread_count", "(", ")", "}" ]
serialized form for channel listing
[ "serialized", "form", "for", "channel", "listing" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L214-L225
zetaops/zengine
zengine/messaging/model.py
Subscriber.create_exchange
def create_exchange(self): """ Creates user's private exchange Actually user's private channel needed to be defined only once, and this should be happened when user first created. But since this has a little performance cost, to be safe we always call it before binding to the channel we currently subscribe """ channel = self._connect_mq() channel.exchange_declare(exchange=self.user.prv_exchange, exchange_type='fanout', durable=True)
python
def create_exchange(self): """ Creates user's private exchange Actually user's private channel needed to be defined only once, and this should be happened when user first created. But since this has a little performance cost, to be safe we always call it before binding to the channel we currently subscribe """ channel = self._connect_mq() channel.exchange_declare(exchange=self.user.prv_exchange, exchange_type='fanout', durable=True)
[ "def", "create_exchange", "(", "self", ")", ":", "channel", "=", "self", ".", "_connect_mq", "(", ")", "channel", ".", "exchange_declare", "(", "exchange", "=", "self", ".", "user", ".", "prv_exchange", ",", "exchange_type", "=", "'fanout'", ",", "durable", "=", "True", ")" ]
Creates user's private exchange Actually user's private channel needed to be defined only once, and this should be happened when user first created. But since this has a little performance cost, to be safe we always call it before binding to the channel we currently subscribe
[ "Creates", "user", "s", "private", "exchange" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L264-L276
zetaops/zengine
zengine/messaging/model.py
Subscriber.bind_to_channel
def bind_to_channel(self): """ Binds (subscribes) users private exchange to channel exchange Automatically called at creation of subscription record. """ if self.channel.code_name != self.user.prv_exchange: channel = self._connect_mq() channel.exchange_bind(source=self.channel.code_name, destination=self.user.prv_exchange)
python
def bind_to_channel(self): """ Binds (subscribes) users private exchange to channel exchange Automatically called at creation of subscription record. """ if self.channel.code_name != self.user.prv_exchange: channel = self._connect_mq() channel.exchange_bind(source=self.channel.code_name, destination=self.user.prv_exchange)
[ "def", "bind_to_channel", "(", "self", ")", ":", "if", "self", ".", "channel", ".", "code_name", "!=", "self", ".", "user", ".", "prv_exchange", ":", "channel", "=", "self", ".", "_connect_mq", "(", ")", "channel", ".", "exchange_bind", "(", "source", "=", "self", ".", "channel", ".", "code_name", ",", "destination", "=", "self", ".", "user", ".", "prv_exchange", ")" ]
Binds (subscribes) users private exchange to channel exchange Automatically called at creation of subscription record.
[ "Binds", "(", "subscribes", ")", "users", "private", "exchange", "to", "channel", "exchange", "Automatically", "called", "at", "creation", "of", "subscription", "record", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L282-L289
zetaops/zengine
zengine/messaging/model.py
Message.serialize
def serialize(self, user=None): """ Serializes message for given user. Note: Should be called before first save(). Otherwise "is_update" will get wrong value. Args: user: User object Returns: Dict. JSON serialization ready dictionary object """ return { 'content': self.body, 'type': self.typ, 'updated_at': self.updated_at, 'timestamp': self.updated_at, 'is_update': not hasattr(self, 'unsaved'), 'attachments': [attachment.serialize() for attachment in self.attachment_set], 'title': self.msg_title, 'url': self.url, 'sender_name': self.sender.full_name, 'sender_key': self.sender.key, 'channel_key': self.channel.key, 'cmd': 'message', 'avatar_url': self.sender.avatar, 'key': self.key, }
python
def serialize(self, user=None): """ Serializes message for given user. Note: Should be called before first save(). Otherwise "is_update" will get wrong value. Args: user: User object Returns: Dict. JSON serialization ready dictionary object """ return { 'content': self.body, 'type': self.typ, 'updated_at': self.updated_at, 'timestamp': self.updated_at, 'is_update': not hasattr(self, 'unsaved'), 'attachments': [attachment.serialize() for attachment in self.attachment_set], 'title': self.msg_title, 'url': self.url, 'sender_name': self.sender.full_name, 'sender_key': self.sender.key, 'channel_key': self.channel.key, 'cmd': 'message', 'avatar_url': self.sender.avatar, 'key': self.key, }
[ "def", "serialize", "(", "self", ",", "user", "=", "None", ")", ":", "return", "{", "'content'", ":", "self", ".", "body", ",", "'type'", ":", "self", ".", "typ", ",", "'updated_at'", ":", "self", ".", "updated_at", ",", "'timestamp'", ":", "self", ".", "updated_at", ",", "'is_update'", ":", "not", "hasattr", "(", "self", ",", "'unsaved'", ")", ",", "'attachments'", ":", "[", "attachment", ".", "serialize", "(", ")", "for", "attachment", "in", "self", ".", "attachment_set", "]", ",", "'title'", ":", "self", ".", "msg_title", ",", "'url'", ":", "self", ".", "url", ",", "'sender_name'", ":", "self", ".", "sender", ".", "full_name", ",", "'sender_key'", ":", "self", ".", "sender", ".", "key", ",", "'channel_key'", ":", "self", ".", "channel", ".", "key", ",", "'cmd'", ":", "'message'", ",", "'avatar_url'", ":", "self", ".", "sender", ".", "avatar", ",", "'key'", ":", "self", ".", "key", ",", "}" ]
Serializes message for given user. Note: Should be called before first save(). Otherwise "is_update" will get wrong value. Args: user: User object Returns: Dict. JSON serialization ready dictionary object
[ "Serializes", "message", "for", "given", "user", "." ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L370-L398
zetaops/zengine
zengine/messaging/model.py
Message._republish
def _republish(self): """ Re-publishes updated message """ mq_channel = self.channel._connect_mq() mq_channel.basic_publish(exchange=self.channel.key, routing_key='', body=json.dumps(self.serialize()))
python
def _republish(self): """ Re-publishes updated message """ mq_channel = self.channel._connect_mq() mq_channel.basic_publish(exchange=self.channel.key, routing_key='', body=json.dumps(self.serialize()))
[ "def", "_republish", "(", "self", ")", ":", "mq_channel", "=", "self", ".", "channel", ".", "_connect_mq", "(", ")", "mq_channel", ".", "basic_publish", "(", "exchange", "=", "self", ".", "channel", ".", "key", ",", "routing_key", "=", "''", ",", "body", "=", "json", ".", "dumps", "(", "self", ".", "serialize", "(", ")", ")", ")" ]
Re-publishes updated message
[ "Re", "-", "publishes", "updated", "message" ]
train
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L404-L410
chrismattmann/nutch-python
nutch/nutch.py
defaultCrawlId
def defaultCrawlId(): """ Provide a reasonable default crawl name using the user name and date """ timestamp = datetime.now().isoformat().replace(':', '_') user = getuser() return '_'.join(('crawl', user, timestamp))
python
def defaultCrawlId(): """ Provide a reasonable default crawl name using the user name and date """ timestamp = datetime.now().isoformat().replace(':', '_') user = getuser() return '_'.join(('crawl', user, timestamp))
[ "def", "defaultCrawlId", "(", ")", ":", "timestamp", "=", "datetime", ".", "now", "(", ")", ".", "isoformat", "(", ")", ".", "replace", "(", "':'", ",", "'_'", ")", "user", "=", "getuser", "(", ")", "return", "'_'", ".", "join", "(", "(", "'crawl'", ",", "user", ",", "timestamp", ")", ")" ]
Provide a reasonable default crawl name using the user name and date
[ "Provide", "a", "reasonable", "default", "crawl", "name", "using", "the", "user", "name", "and", "date" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L91-L98
chrismattmann/nutch-python
nutch/nutch.py
main
def main(argv=None): """Run Nutch command using REST API.""" global Verbose, Mock if argv is None: argv = sys.argv if len(argv) < 5: die('Bad args') try: opts, argv = getopt.getopt(argv[1:], 'hs:p:mv', ['help', 'server=', 'port=', 'mock', 'verbose']) except getopt.GetoptError as err: # print help information and exit: print(err) # will print something like "option -a not recognized" die() serverEndpoint = DefaultServerEndpoint # TODO: Fix this for opt, val in opts: if opt in ('-h', '--help'): echo2(USAGE); sys.exit() elif opt in ('-s', '--server'): serverEndpoint = val elif opt in ('-p', '--port'): serverEndpoint = 'http://localhost:%s' % val elif opt in ('-m', '--mock'): Mock = 1 elif opt in ('-v', '--verbose'): Verbose = 1 else: die(USAGE) cmd = argv[0] crawlId = argv[1] confId = argv[2] urlDir = argv[3] args = {} if len(argv) > 4: args = eval(argv[4]) nt = Nutch(crawlId, confId, serverEndpoint, urlDir) nt.Jobs().create(cmd, **args)
python
def main(argv=None): """Run Nutch command using REST API.""" global Verbose, Mock if argv is None: argv = sys.argv if len(argv) < 5: die('Bad args') try: opts, argv = getopt.getopt(argv[1:], 'hs:p:mv', ['help', 'server=', 'port=', 'mock', 'verbose']) except getopt.GetoptError as err: # print help information and exit: print(err) # will print something like "option -a not recognized" die() serverEndpoint = DefaultServerEndpoint # TODO: Fix this for opt, val in opts: if opt in ('-h', '--help'): echo2(USAGE); sys.exit() elif opt in ('-s', '--server'): serverEndpoint = val elif opt in ('-p', '--port'): serverEndpoint = 'http://localhost:%s' % val elif opt in ('-m', '--mock'): Mock = 1 elif opt in ('-v', '--verbose'): Verbose = 1 else: die(USAGE) cmd = argv[0] crawlId = argv[1] confId = argv[2] urlDir = argv[3] args = {} if len(argv) > 4: args = eval(argv[4]) nt = Nutch(crawlId, confId, serverEndpoint, urlDir) nt.Jobs().create(cmd, **args)
[ "def", "main", "(", "argv", "=", "None", ")", ":", "global", "Verbose", ",", "Mock", "if", "argv", "is", "None", ":", "argv", "=", "sys", ".", "argv", "if", "len", "(", "argv", ")", "<", "5", ":", "die", "(", "'Bad args'", ")", "try", ":", "opts", ",", "argv", "=", "getopt", ".", "getopt", "(", "argv", "[", "1", ":", "]", ",", "'hs:p:mv'", ",", "[", "'help'", ",", "'server='", ",", "'port='", ",", "'mock'", ",", "'verbose'", "]", ")", "except", "getopt", ".", "GetoptError", "as", "err", ":", "# print help information and exit:", "print", "(", "err", ")", "# will print something like \"option -a not recognized\"", "die", "(", ")", "serverEndpoint", "=", "DefaultServerEndpoint", "# TODO: Fix this", "for", "opt", ",", "val", "in", "opts", ":", "if", "opt", "in", "(", "'-h'", ",", "'--help'", ")", ":", "echo2", "(", "USAGE", ")", "sys", ".", "exit", "(", ")", "elif", "opt", "in", "(", "'-s'", ",", "'--server'", ")", ":", "serverEndpoint", "=", "val", "elif", "opt", "in", "(", "'-p'", ",", "'--port'", ")", ":", "serverEndpoint", "=", "'http://localhost:%s'", "%", "val", "elif", "opt", "in", "(", "'-m'", ",", "'--mock'", ")", ":", "Mock", "=", "1", "elif", "opt", "in", "(", "'-v'", ",", "'--verbose'", ")", ":", "Verbose", "=", "1", "else", ":", "die", "(", "USAGE", ")", "cmd", "=", "argv", "[", "0", "]", "crawlId", "=", "argv", "[", "1", "]", "confId", "=", "argv", "[", "2", "]", "urlDir", "=", "argv", "[", "3", "]", "args", "=", "{", "}", "if", "len", "(", "argv", ")", ">", "4", ":", "args", "=", "eval", "(", "argv", "[", "4", "]", ")", "nt", "=", "Nutch", "(", "crawlId", ",", "confId", ",", "serverEndpoint", ",", "urlDir", ")", "nt", ".", "Jobs", "(", ")", ".", "create", "(", "cmd", ",", "*", "*", "args", ")" ]
Run Nutch command using REST API.
[ "Run", "Nutch", "command", "using", "REST", "API", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L716-L749
chrismattmann/nutch-python
nutch/nutch.py
Server.call
def call(self, verb, servicePath, data=None, headers=None, forceText=False, sendJson=True): """Call the Nutch Server, do some error checking, and return the response. :param verb: One of nutch.RequestVerbs :param servicePath: path component of URL to append to endpoint, e.g. '/config' :param data: Data to attach to this request :param headers: headers to attach to this request, default are JsonAcceptHeader :param forceText: don't trust the response headers and just get the text :param sendJson: Whether to treat attached data as JSON or not """ default_data = {} if sendJson else "" data = data if data else default_data headers = headers if headers else JsonAcceptHeader.copy() if not sendJson: headers.update(TextSendHeader) if verb not in RequestVerbs: die('Server call verb must be one of %s' % str(RequestVerbs.keys())) if Verbose: echo2("%s Endpoint:" % verb.upper(), servicePath) echo2("%s Request data:" % verb.upper(), data) echo2("%s Request headers:" % verb.upper(), headers) verbFn = RequestVerbs[verb] if sendJson: resp = verbFn(self.serverEndpoint + servicePath, json=data, headers=headers) else: resp = verbFn(self.serverEndpoint + servicePath, data=data, headers=headers) if Verbose: echo2("Response headers:", resp.headers) echo2("Response status:", resp.status_code) if resp.status_code != 200: if self.raiseErrors: error = NutchException("Unexpected server response: %d" % resp.status_code) error.status_code = resp.status_code raise error else: warn('Nutch server returned status:', resp.status_code) if forceText or 'content-type' not in resp.headers or resp.headers['content-type'] == 'text/plain': if Verbose: echo2("Response text:", resp.text) return resp.text content_type = resp.headers['content-type'] if content_type == 'application/json' and not forceText: if Verbose: echo2("Response JSON:", resp.json()) return resp.json() else: die('Did not understand server response: %s' % resp.headers)
python
def call(self, verb, servicePath, data=None, headers=None, forceText=False, sendJson=True): """Call the Nutch Server, do some error checking, and return the response. :param verb: One of nutch.RequestVerbs :param servicePath: path component of URL to append to endpoint, e.g. '/config' :param data: Data to attach to this request :param headers: headers to attach to this request, default are JsonAcceptHeader :param forceText: don't trust the response headers and just get the text :param sendJson: Whether to treat attached data as JSON or not """ default_data = {} if sendJson else "" data = data if data else default_data headers = headers if headers else JsonAcceptHeader.copy() if not sendJson: headers.update(TextSendHeader) if verb not in RequestVerbs: die('Server call verb must be one of %s' % str(RequestVerbs.keys())) if Verbose: echo2("%s Endpoint:" % verb.upper(), servicePath) echo2("%s Request data:" % verb.upper(), data) echo2("%s Request headers:" % verb.upper(), headers) verbFn = RequestVerbs[verb] if sendJson: resp = verbFn(self.serverEndpoint + servicePath, json=data, headers=headers) else: resp = verbFn(self.serverEndpoint + servicePath, data=data, headers=headers) if Verbose: echo2("Response headers:", resp.headers) echo2("Response status:", resp.status_code) if resp.status_code != 200: if self.raiseErrors: error = NutchException("Unexpected server response: %d" % resp.status_code) error.status_code = resp.status_code raise error else: warn('Nutch server returned status:', resp.status_code) if forceText or 'content-type' not in resp.headers or resp.headers['content-type'] == 'text/plain': if Verbose: echo2("Response text:", resp.text) return resp.text content_type = resp.headers['content-type'] if content_type == 'application/json' and not forceText: if Verbose: echo2("Response JSON:", resp.json()) return resp.json() else: die('Did not understand server response: %s' % resp.headers)
[ "def", "call", "(", "self", ",", "verb", ",", "servicePath", ",", "data", "=", "None", ",", "headers", "=", "None", ",", "forceText", "=", "False", ",", "sendJson", "=", "True", ")", ":", "default_data", "=", "{", "}", "if", "sendJson", "else", "\"\"", "data", "=", "data", "if", "data", "else", "default_data", "headers", "=", "headers", "if", "headers", "else", "JsonAcceptHeader", ".", "copy", "(", ")", "if", "not", "sendJson", ":", "headers", ".", "update", "(", "TextSendHeader", ")", "if", "verb", "not", "in", "RequestVerbs", ":", "die", "(", "'Server call verb must be one of %s'", "%", "str", "(", "RequestVerbs", ".", "keys", "(", ")", ")", ")", "if", "Verbose", ":", "echo2", "(", "\"%s Endpoint:\"", "%", "verb", ".", "upper", "(", ")", ",", "servicePath", ")", "echo2", "(", "\"%s Request data:\"", "%", "verb", ".", "upper", "(", ")", ",", "data", ")", "echo2", "(", "\"%s Request headers:\"", "%", "verb", ".", "upper", "(", ")", ",", "headers", ")", "verbFn", "=", "RequestVerbs", "[", "verb", "]", "if", "sendJson", ":", "resp", "=", "verbFn", "(", "self", ".", "serverEndpoint", "+", "servicePath", ",", "json", "=", "data", ",", "headers", "=", "headers", ")", "else", ":", "resp", "=", "verbFn", "(", "self", ".", "serverEndpoint", "+", "servicePath", ",", "data", "=", "data", ",", "headers", "=", "headers", ")", "if", "Verbose", ":", "echo2", "(", "\"Response headers:\"", ",", "resp", ".", "headers", ")", "echo2", "(", "\"Response status:\"", ",", "resp", ".", "status_code", ")", "if", "resp", ".", "status_code", "!=", "200", ":", "if", "self", ".", "raiseErrors", ":", "error", "=", "NutchException", "(", "\"Unexpected server response: %d\"", "%", "resp", ".", "status_code", ")", "error", ".", "status_code", "=", "resp", ".", "status_code", "raise", "error", "else", ":", "warn", "(", "'Nutch server returned status:'", ",", "resp", ".", "status_code", ")", "if", "forceText", "or", "'content-type'", "not", "in", "resp", ".", "headers", "or", "resp", ".", "headers", "[", "'content-type'", "]", "==", "'text/plain'", ":", "if", "Verbose", ":", "echo2", "(", "\"Response text:\"", ",", "resp", ".", "text", ")", "return", "resp", ".", "text", "content_type", "=", "resp", ".", "headers", "[", "'content-type'", "]", "if", "content_type", "==", "'application/json'", "and", "not", "forceText", ":", "if", "Verbose", ":", "echo2", "(", "\"Response JSON:\"", ",", "resp", ".", "json", "(", ")", ")", "return", "resp", ".", "json", "(", ")", "else", ":", "die", "(", "'Did not understand server response: %s'", "%", "resp", ".", "headers", ")" ]
Call the Nutch Server, do some error checking, and return the response. :param verb: One of nutch.RequestVerbs :param servicePath: path component of URL to append to endpoint, e.g. '/config' :param data: Data to attach to this request :param headers: headers to attach to this request, default are JsonAcceptHeader :param forceText: don't trust the response headers and just get the text :param sendJson: Whether to treat attached data as JSON or not
[ "Call", "the", "Nutch", "Server", "do", "some", "error", "checking", "and", "return", "the", "response", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L117-L170
chrismattmann/nutch-python
nutch/nutch.py
ConfigClient.create
def create(self, cid, configData): """ Create a new named (cid) configuration from a parameter dictionary (config_data). """ configArgs = {'configId': cid, 'params': configData, 'force': True} cid = self.server.call('post', "/config/create", configArgs, forceText=True, headers=TextAcceptHeader) new_config = Config(cid, self.server) return new_config
python
def create(self, cid, configData): """ Create a new named (cid) configuration from a parameter dictionary (config_data). """ configArgs = {'configId': cid, 'params': configData, 'force': True} cid = self.server.call('post', "/config/create", configArgs, forceText=True, headers=TextAcceptHeader) new_config = Config(cid, self.server) return new_config
[ "def", "create", "(", "self", ",", "cid", ",", "configData", ")", ":", "configArgs", "=", "{", "'configId'", ":", "cid", ",", "'params'", ":", "configData", ",", "'force'", ":", "True", "}", "cid", "=", "self", ".", "server", ".", "call", "(", "'post'", ",", "\"/config/create\"", ",", "configArgs", ",", "forceText", "=", "True", ",", "headers", "=", "TextAcceptHeader", ")", "new_config", "=", "Config", "(", "cid", ",", "self", ".", "server", ")", "return", "new_config" ]
Create a new named (cid) configuration from a parameter dictionary (config_data).
[ "Create", "a", "new", "named", "(", "cid", ")", "configuration", "from", "a", "parameter", "dictionary", "(", "config_data", ")", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L278-L285
chrismattmann/nutch-python
nutch/nutch.py
JobClient.list
def list(self, allJobs=False): """ Return list of jobs at this endpoint. Call get(allJobs=True) to see all jobs, not just the ones managed by this Client """ jobs = self.server.call('get', '/job') return [Job(job['id'], self.server) for job in jobs if allJobs or self._job_owned(job)]
python
def list(self, allJobs=False): """ Return list of jobs at this endpoint. Call get(allJobs=True) to see all jobs, not just the ones managed by this Client """ jobs = self.server.call('get', '/job') return [Job(job['id'], self.server) for job in jobs if allJobs or self._job_owned(job)]
[ "def", "list", "(", "self", ",", "allJobs", "=", "False", ")", ":", "jobs", "=", "self", ".", "server", ".", "call", "(", "'get'", ",", "'/job'", ")", "return", "[", "Job", "(", "job", "[", "'id'", "]", ",", "self", ".", "server", ")", "for", "job", "in", "jobs", "if", "allJobs", "or", "self", ".", "_job_owned", "(", "job", ")", "]" ]
Return list of jobs at this endpoint. Call get(allJobs=True) to see all jobs, not just the ones managed by this Client
[ "Return", "list", "of", "jobs", "at", "this", "endpoint", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L337-L346
chrismattmann/nutch-python
nutch/nutch.py
JobClient.create
def create(self, command, **args): """ Create a job given a command :param command: Nutch command, one of nutch.LegalJobs :param args: Additional arguments to pass to the job :return: The created Job """ command = command.upper() if command not in LegalJobs: warn('Nutch command must be one of: %s' % ', '.join(LegalJobs)) else: echo2('Starting %s job with args %s' % (command, str(args))) parameters = self.parameters.copy() parameters['type'] = command parameters['crawlId'] = self.crawlId parameters['confId'] = self.confId parameters['args'].update(args) job_info = self.server.call('post', "/job/create", parameters, JsonAcceptHeader) job = Job(job_info['id'], self.server) return job
python
def create(self, command, **args): """ Create a job given a command :param command: Nutch command, one of nutch.LegalJobs :param args: Additional arguments to pass to the job :return: The created Job """ command = command.upper() if command not in LegalJobs: warn('Nutch command must be one of: %s' % ', '.join(LegalJobs)) else: echo2('Starting %s job with args %s' % (command, str(args))) parameters = self.parameters.copy() parameters['type'] = command parameters['crawlId'] = self.crawlId parameters['confId'] = self.confId parameters['args'].update(args) job_info = self.server.call('post', "/job/create", parameters, JsonAcceptHeader) job = Job(job_info['id'], self.server) return job
[ "def", "create", "(", "self", ",", "command", ",", "*", "*", "args", ")", ":", "command", "=", "command", ".", "upper", "(", ")", "if", "command", "not", "in", "LegalJobs", ":", "warn", "(", "'Nutch command must be one of: %s'", "%", "', '", ".", "join", "(", "LegalJobs", ")", ")", "else", ":", "echo2", "(", "'Starting %s job with args %s'", "%", "(", "command", ",", "str", "(", "args", ")", ")", ")", "parameters", "=", "self", ".", "parameters", ".", "copy", "(", ")", "parameters", "[", "'type'", "]", "=", "command", "parameters", "[", "'crawlId'", "]", "=", "self", ".", "crawlId", "parameters", "[", "'confId'", "]", "=", "self", ".", "confId", "parameters", "[", "'args'", "]", ".", "update", "(", "args", ")", "job_info", "=", "self", ".", "server", ".", "call", "(", "'post'", ",", "\"/job/create\"", ",", "parameters", ",", "JsonAcceptHeader", ")", "job", "=", "Job", "(", "job_info", "[", "'id'", "]", ",", "self", ".", "server", ")", "return", "job" ]
Create a job given a command :param command: Nutch command, one of nutch.LegalJobs :param args: Additional arguments to pass to the job :return: The created Job
[ "Create", "a", "job", "given", "a", "command", ":", "param", "command", ":", "Nutch", "command", "one", "of", "nutch", ".", "LegalJobs", ":", "param", "args", ":", "Additional", "arguments", "to", "pass", "to", "the", "job", ":", "return", ":", "The", "created", "Job" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L348-L370
chrismattmann/nutch-python
nutch/nutch.py
JobClient.inject
def inject(self, seed=None, urlDir=None, **args): """ :param seed: A Seed object (this or urlDir must be specified) :param urlDir: The directory on the server containing the seed list (this or urlDir must be specified) :param args: Extra arguments for the job :return: a created Job object """ if seed: if urlDir and urlDir != seed.seedPath: raise NutchException("Can't specify both seed and urlDir") urlDir = seed.seedPath elif urlDir: pass else: raise NutchException("Must specify seed or urlDir") args['url_dir'] = urlDir return self.create('INJECT', **args)
python
def inject(self, seed=None, urlDir=None, **args): """ :param seed: A Seed object (this or urlDir must be specified) :param urlDir: The directory on the server containing the seed list (this or urlDir must be specified) :param args: Extra arguments for the job :return: a created Job object """ if seed: if urlDir and urlDir != seed.seedPath: raise NutchException("Can't specify both seed and urlDir") urlDir = seed.seedPath elif urlDir: pass else: raise NutchException("Must specify seed or urlDir") args['url_dir'] = urlDir return self.create('INJECT', **args)
[ "def", "inject", "(", "self", ",", "seed", "=", "None", ",", "urlDir", "=", "None", ",", "*", "*", "args", ")", ":", "if", "seed", ":", "if", "urlDir", "and", "urlDir", "!=", "seed", ".", "seedPath", ":", "raise", "NutchException", "(", "\"Can't specify both seed and urlDir\"", ")", "urlDir", "=", "seed", ".", "seedPath", "elif", "urlDir", ":", "pass", "else", ":", "raise", "NutchException", "(", "\"Must specify seed or urlDir\"", ")", "args", "[", "'url_dir'", "]", "=", "urlDir", "return", "self", ".", "create", "(", "'INJECT'", ",", "*", "*", "args", ")" ]
:param seed: A Seed object (this or urlDir must be specified) :param urlDir: The directory on the server containing the seed list (this or urlDir must be specified) :param args: Extra arguments for the job :return: a created Job object
[ ":", "param", "seed", ":", "A", "Seed", "object", "(", "this", "or", "urlDir", "must", "be", "specified", ")", ":", "param", "urlDir", ":", "The", "directory", "on", "the", "server", "containing", "the", "seed", "list", "(", "this", "or", "urlDir", "must", "be", "specified", ")", ":", "param", "args", ":", "Extra", "arguments", "for", "the", "job", ":", "return", ":", "a", "created", "Job", "object" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L374-L391
chrismattmann/nutch-python
nutch/nutch.py
SeedClient.create
def create(self, sid, seedList): """ Create a new named (sid) Seed from a list of seed URLs :param sid: the name to assign to the new seed list :param seedList: the list of seeds to use :return: the created Seed object """ seedUrl = lambda uid, url: {"id": uid, "url": url} if not isinstance(seedList,tuple): seedList = (seedList,) seedListData = { "id": "12345", "name": sid, "seedUrls": [seedUrl(uid, url) for uid, url in enumerate(seedList)] } # As per resolution of https://issues.apache.org/jira/browse/NUTCH-2123 seedPath = self.server.call('post', "/seed/create", seedListData, TextAcceptHeader) new_seed = Seed(sid, seedPath, self.server) return new_seed
python
def create(self, sid, seedList): """ Create a new named (sid) Seed from a list of seed URLs :param sid: the name to assign to the new seed list :param seedList: the list of seeds to use :return: the created Seed object """ seedUrl = lambda uid, url: {"id": uid, "url": url} if not isinstance(seedList,tuple): seedList = (seedList,) seedListData = { "id": "12345", "name": sid, "seedUrls": [seedUrl(uid, url) for uid, url in enumerate(seedList)] } # As per resolution of https://issues.apache.org/jira/browse/NUTCH-2123 seedPath = self.server.call('post', "/seed/create", seedListData, TextAcceptHeader) new_seed = Seed(sid, seedPath, self.server) return new_seed
[ "def", "create", "(", "self", ",", "sid", ",", "seedList", ")", ":", "seedUrl", "=", "lambda", "uid", ",", "url", ":", "{", "\"id\"", ":", "uid", ",", "\"url\"", ":", "url", "}", "if", "not", "isinstance", "(", "seedList", ",", "tuple", ")", ":", "seedList", "=", "(", "seedList", ",", ")", "seedListData", "=", "{", "\"id\"", ":", "\"12345\"", ",", "\"name\"", ":", "sid", ",", "\"seedUrls\"", ":", "[", "seedUrl", "(", "uid", ",", "url", ")", "for", "uid", ",", "url", "in", "enumerate", "(", "seedList", ")", "]", "}", "# As per resolution of https://issues.apache.org/jira/browse/NUTCH-2123", "seedPath", "=", "self", ".", "server", ".", "call", "(", "'post'", ",", "\"/seed/create\"", ",", "seedListData", ",", "TextAcceptHeader", ")", "new_seed", "=", "Seed", "(", "sid", ",", "seedPath", ",", "self", ".", "server", ")", "return", "new_seed" ]
Create a new named (sid) Seed from a list of seed URLs :param sid: the name to assign to the new seed list :param seedList: the list of seeds to use :return: the created Seed object
[ "Create", "a", "new", "named", "(", "sid", ")", "Seed", "from", "a", "list", "of", "seed", "URLs" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L419-L442
chrismattmann/nutch-python
nutch/nutch.py
SeedClient.createFromFile
def createFromFile(self, sid, filename): """ Create a new named (sid) Seed from a file containing URLs It's assumed URLs are whitespace seperated. :param sid: the name to assign to the new seed list :param filename: the name of the file that contains URLs :return: the created Seed object """ urls = [] with open(filename) as f: for line in f: for url in line.split(): urls.append(url) return self.create(sid, tuple(urls))
python
def createFromFile(self, sid, filename): """ Create a new named (sid) Seed from a file containing URLs It's assumed URLs are whitespace seperated. :param sid: the name to assign to the new seed list :param filename: the name of the file that contains URLs :return: the created Seed object """ urls = [] with open(filename) as f: for line in f: for url in line.split(): urls.append(url) return self.create(sid, tuple(urls))
[ "def", "createFromFile", "(", "self", ",", "sid", ",", "filename", ")", ":", "urls", "=", "[", "]", "with", "open", "(", "filename", ")", "as", "f", ":", "for", "line", "in", "f", ":", "for", "url", "in", "line", ".", "split", "(", ")", ":", "urls", ".", "append", "(", "url", ")", "return", "self", ".", "create", "(", "sid", ",", "tuple", "(", "urls", ")", ")" ]
Create a new named (sid) Seed from a file containing URLs It's assumed URLs are whitespace seperated. :param sid: the name to assign to the new seed list :param filename: the name of the file that contains URLs :return: the created Seed object
[ "Create", "a", "new", "named", "(", "sid", ")", "Seed", "from", "a", "file", "containing", "URLs", "It", "s", "assumed", "URLs", "are", "whitespace", "seperated", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L444-L460
chrismattmann/nutch-python
nutch/nutch.py
CrawlClient._nextJob
def _nextJob(self, job, nextRound=True): """ Given a completed job, start the next job in the round, or return None :param nextRound: whether to start jobs from the next round if the current round is completed. :return: the newly started Job, or None if no job was started """ jobInfo = job.info() assert jobInfo['state'] == 'FINISHED' roundEnd = False if jobInfo['type'] == 'INJECT': nextCommand = 'GENERATE' elif jobInfo['type'] == 'GENERATE': nextCommand = 'FETCH' elif jobInfo['type'] == 'FETCH': nextCommand = 'PARSE' elif jobInfo['type'] == 'PARSE': nextCommand = 'UPDATEDB' elif jobInfo['type'] == 'UPDATEDB': nextCommand = 'INVERTLINKS' elif jobInfo['type'] == 'INVERTLINKS': nextCommand = 'DEDUP' elif jobInfo['type'] == 'DEDUP': if self.enable_index: nextCommand = 'INDEX' else: roundEnd = True elif jobInfo['type'] == 'INDEX': roundEnd = True else: raise NutchException("Unrecognized job type {}".format(jobInfo['type'])) if roundEnd: if nextRound and self.currentRound < self.totalRounds: nextCommand = 'GENERATE' self.currentRound += 1 else: return None return self.jobClient.create(nextCommand)
python
def _nextJob(self, job, nextRound=True): """ Given a completed job, start the next job in the round, or return None :param nextRound: whether to start jobs from the next round if the current round is completed. :return: the newly started Job, or None if no job was started """ jobInfo = job.info() assert jobInfo['state'] == 'FINISHED' roundEnd = False if jobInfo['type'] == 'INJECT': nextCommand = 'GENERATE' elif jobInfo['type'] == 'GENERATE': nextCommand = 'FETCH' elif jobInfo['type'] == 'FETCH': nextCommand = 'PARSE' elif jobInfo['type'] == 'PARSE': nextCommand = 'UPDATEDB' elif jobInfo['type'] == 'UPDATEDB': nextCommand = 'INVERTLINKS' elif jobInfo['type'] == 'INVERTLINKS': nextCommand = 'DEDUP' elif jobInfo['type'] == 'DEDUP': if self.enable_index: nextCommand = 'INDEX' else: roundEnd = True elif jobInfo['type'] == 'INDEX': roundEnd = True else: raise NutchException("Unrecognized job type {}".format(jobInfo['type'])) if roundEnd: if nextRound and self.currentRound < self.totalRounds: nextCommand = 'GENERATE' self.currentRound += 1 else: return None return self.jobClient.create(nextCommand)
[ "def", "_nextJob", "(", "self", ",", "job", ",", "nextRound", "=", "True", ")", ":", "jobInfo", "=", "job", ".", "info", "(", ")", "assert", "jobInfo", "[", "'state'", "]", "==", "'FINISHED'", "roundEnd", "=", "False", "if", "jobInfo", "[", "'type'", "]", "==", "'INJECT'", ":", "nextCommand", "=", "'GENERATE'", "elif", "jobInfo", "[", "'type'", "]", "==", "'GENERATE'", ":", "nextCommand", "=", "'FETCH'", "elif", "jobInfo", "[", "'type'", "]", "==", "'FETCH'", ":", "nextCommand", "=", "'PARSE'", "elif", "jobInfo", "[", "'type'", "]", "==", "'PARSE'", ":", "nextCommand", "=", "'UPDATEDB'", "elif", "jobInfo", "[", "'type'", "]", "==", "'UPDATEDB'", ":", "nextCommand", "=", "'INVERTLINKS'", "elif", "jobInfo", "[", "'type'", "]", "==", "'INVERTLINKS'", ":", "nextCommand", "=", "'DEDUP'", "elif", "jobInfo", "[", "'type'", "]", "==", "'DEDUP'", ":", "if", "self", ".", "enable_index", ":", "nextCommand", "=", "'INDEX'", "else", ":", "roundEnd", "=", "True", "elif", "jobInfo", "[", "'type'", "]", "==", "'INDEX'", ":", "roundEnd", "=", "True", "else", ":", "raise", "NutchException", "(", "\"Unrecognized job type {}\"", ".", "format", "(", "jobInfo", "[", "'type'", "]", ")", ")", "if", "roundEnd", ":", "if", "nextRound", "and", "self", ".", "currentRound", "<", "self", ".", "totalRounds", ":", "nextCommand", "=", "'GENERATE'", "self", ".", "currentRound", "+=", "1", "else", ":", "return", "None", "return", "self", ".", "jobClient", ".", "create", "(", "nextCommand", ")" ]
Given a completed job, start the next job in the round, or return None :param nextRound: whether to start jobs from the next round if the current round is completed. :return: the newly started Job, or None if no job was started
[ "Given", "a", "completed", "job", "start", "the", "next", "job", "in", "the", "round", "or", "return", "None" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L492-L533
chrismattmann/nutch-python
nutch/nutch.py
CrawlClient.progress
def progress(self, nextRound=True): """ Check the status of the current job, activate the next job if it's finished, and return the active job If the current job has failed, a NutchCrawlException will be raised with no jobs attached. :param nextRound: whether to start jobs from the next round if the current job/round is completed. :return: the currently running Job, or None if no jobs are running. """ currentJob = self.currentJob if currentJob is None: return currentJob jobInfo = currentJob.info() if jobInfo['state'] == 'RUNNING': return currentJob elif jobInfo['state'] == 'FINISHED': nextJob = self._nextJob(currentJob, nextRound) self.currentJob = nextJob return nextJob else: error = NutchCrawlException("Unexpected job state: {}".format(jobInfo['state'])) error.current_job = currentJob raise NutchCrawlException
python
def progress(self, nextRound=True): """ Check the status of the current job, activate the next job if it's finished, and return the active job If the current job has failed, a NutchCrawlException will be raised with no jobs attached. :param nextRound: whether to start jobs from the next round if the current job/round is completed. :return: the currently running Job, or None if no jobs are running. """ currentJob = self.currentJob if currentJob is None: return currentJob jobInfo = currentJob.info() if jobInfo['state'] == 'RUNNING': return currentJob elif jobInfo['state'] == 'FINISHED': nextJob = self._nextJob(currentJob, nextRound) self.currentJob = nextJob return nextJob else: error = NutchCrawlException("Unexpected job state: {}".format(jobInfo['state'])) error.current_job = currentJob raise NutchCrawlException
[ "def", "progress", "(", "self", ",", "nextRound", "=", "True", ")", ":", "currentJob", "=", "self", ".", "currentJob", "if", "currentJob", "is", "None", ":", "return", "currentJob", "jobInfo", "=", "currentJob", ".", "info", "(", ")", "if", "jobInfo", "[", "'state'", "]", "==", "'RUNNING'", ":", "return", "currentJob", "elif", "jobInfo", "[", "'state'", "]", "==", "'FINISHED'", ":", "nextJob", "=", "self", ".", "_nextJob", "(", "currentJob", ",", "nextRound", ")", "self", ".", "currentJob", "=", "nextJob", "return", "nextJob", "else", ":", "error", "=", "NutchCrawlException", "(", "\"Unexpected job state: {}\"", ".", "format", "(", "jobInfo", "[", "'state'", "]", ")", ")", "error", ".", "current_job", "=", "currentJob", "raise", "NutchCrawlException" ]
Check the status of the current job, activate the next job if it's finished, and return the active job If the current job has failed, a NutchCrawlException will be raised with no jobs attached. :param nextRound: whether to start jobs from the next round if the current job/round is completed. :return: the currently running Job, or None if no jobs are running.
[ "Check", "the", "status", "of", "the", "current", "job", "activate", "the", "next", "job", "if", "it", "s", "finished", "and", "return", "the", "active", "job" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L535-L560
chrismattmann/nutch-python
nutch/nutch.py
CrawlClient.nextRound
def nextRound(self): """ Execute all jobs in the current round and return when they have finished. If a job fails, a NutchCrawlException will be raised, with all completed jobs from this round attached to the exception. :return: a list of all completed Jobs """ finishedJobs = [] if self.currentJob is None: self.currentJob = self.jobClient.create('GENERATE') activeJob = self.progress(nextRound=False) while activeJob: oldJob = activeJob activeJob = self.progress(nextRound=False) # updates self.currentJob if oldJob and oldJob != activeJob: finishedJobs.append(oldJob) sleep(self.sleepTime) self.currentRound += 1 return finishedJobs
python
def nextRound(self): """ Execute all jobs in the current round and return when they have finished. If a job fails, a NutchCrawlException will be raised, with all completed jobs from this round attached to the exception. :return: a list of all completed Jobs """ finishedJobs = [] if self.currentJob is None: self.currentJob = self.jobClient.create('GENERATE') activeJob = self.progress(nextRound=False) while activeJob: oldJob = activeJob activeJob = self.progress(nextRound=False) # updates self.currentJob if oldJob and oldJob != activeJob: finishedJobs.append(oldJob) sleep(self.sleepTime) self.currentRound += 1 return finishedJobs
[ "def", "nextRound", "(", "self", ")", ":", "finishedJobs", "=", "[", "]", "if", "self", ".", "currentJob", "is", "None", ":", "self", ".", "currentJob", "=", "self", ".", "jobClient", ".", "create", "(", "'GENERATE'", ")", "activeJob", "=", "self", ".", "progress", "(", "nextRound", "=", "False", ")", "while", "activeJob", ":", "oldJob", "=", "activeJob", "activeJob", "=", "self", ".", "progress", "(", "nextRound", "=", "False", ")", "# updates self.currentJob", "if", "oldJob", "and", "oldJob", "!=", "activeJob", ":", "finishedJobs", ".", "append", "(", "oldJob", ")", "sleep", "(", "self", ".", "sleepTime", ")", "self", ".", "currentRound", "+=", "1", "return", "finishedJobs" ]
Execute all jobs in the current round and return when they have finished. If a job fails, a NutchCrawlException will be raised, with all completed jobs from this round attached to the exception. :return: a list of all completed Jobs
[ "Execute", "all", "jobs", "in", "the", "current", "round", "and", "return", "when", "they", "have", "finished", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L573-L595
chrismattmann/nutch-python
nutch/nutch.py
CrawlClient.waitAll
def waitAll(self): """ Execute all queued rounds and return when they have finished. If a job fails, a NutchCrawlException will be raised, with all completed jobs attached to the exception :return: a list of jobs completed for each round, organized by round (list-of-lists) """ finishedRounds = [self.nextRound()] while self.currentRound < self.totalRounds: finishedRounds.append(self.nextRound()) return finishedRounds
python
def waitAll(self): """ Execute all queued rounds and return when they have finished. If a job fails, a NutchCrawlException will be raised, with all completed jobs attached to the exception :return: a list of jobs completed for each round, organized by round (list-of-lists) """ finishedRounds = [self.nextRound()] while self.currentRound < self.totalRounds: finishedRounds.append(self.nextRound()) return finishedRounds
[ "def", "waitAll", "(", "self", ")", ":", "finishedRounds", "=", "[", "self", ".", "nextRound", "(", ")", "]", "while", "self", ".", "currentRound", "<", "self", ".", "totalRounds", ":", "finishedRounds", ".", "append", "(", "self", ".", "nextRound", "(", ")", ")", "return", "finishedRounds" ]
Execute all queued rounds and return when they have finished. If a job fails, a NutchCrawlException will be raised, with all completed jobs attached to the exception :return: a list of jobs completed for each round, organized by round (list-of-lists)
[ "Execute", "all", "queued", "rounds", "and", "return", "when", "they", "have", "finished", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L597-L612
chrismattmann/nutch-python
nutch/nutch.py
Nutch.Jobs
def Jobs(self, crawlId=None): """ Create a JobClient for listing and creating jobs. The JobClient inherits the confId from the Nutch client. :param crawlId: crawlIds to use for this client. If not provided, will be generated by nutch.defaultCrawlId() :return: a JobClient """ crawlId = crawlId if crawlId else defaultCrawlId() return JobClient(self.server, crawlId, self.confId)
python
def Jobs(self, crawlId=None): """ Create a JobClient for listing and creating jobs. The JobClient inherits the confId from the Nutch client. :param crawlId: crawlIds to use for this client. If not provided, will be generated by nutch.defaultCrawlId() :return: a JobClient """ crawlId = crawlId if crawlId else defaultCrawlId() return JobClient(self.server, crawlId, self.confId)
[ "def", "Jobs", "(", "self", ",", "crawlId", "=", "None", ")", ":", "crawlId", "=", "crawlId", "if", "crawlId", "else", "defaultCrawlId", "(", ")", "return", "JobClient", "(", "self", ".", "server", ",", "crawlId", ",", "self", ".", "confId", ")" ]
Create a JobClient for listing and creating jobs. The JobClient inherits the confId from the Nutch client. :param crawlId: crawlIds to use for this client. If not provided, will be generated by nutch.defaultCrawlId() :return: a JobClient
[ "Create", "a", "JobClient", "for", "listing", "and", "creating", "jobs", ".", "The", "JobClient", "inherits", "the", "confId", "from", "the", "Nutch", "client", "." ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L656-L666
chrismattmann/nutch-python
nutch/nutch.py
Nutch.Crawl
def Crawl(self, seed, seedClient=None, jobClient=None, rounds=1, index=True): """ Launch a crawl using the given seed :param seed: Type (Seed or SeedList) - used for crawl :param seedClient: if a SeedList is given, the SeedClient to upload, if None a default will be created :param jobClient: the JobClient to be used, if None a default will be created :param rounds: the number of rounds in the crawl :return: a CrawlClient to monitor and control the crawl """ if seedClient is None: seedClient = self.Seeds() if jobClient is None: jobClient = self.Jobs() if type(seed) != Seed: seed = seedClient.create(jobClient.crawlId + '_seeds', seed) return CrawlClient(self.server, seed, jobClient, rounds, index)
python
def Crawl(self, seed, seedClient=None, jobClient=None, rounds=1, index=True): """ Launch a crawl using the given seed :param seed: Type (Seed or SeedList) - used for crawl :param seedClient: if a SeedList is given, the SeedClient to upload, if None a default will be created :param jobClient: the JobClient to be used, if None a default will be created :param rounds: the number of rounds in the crawl :return: a CrawlClient to monitor and control the crawl """ if seedClient is None: seedClient = self.Seeds() if jobClient is None: jobClient = self.Jobs() if type(seed) != Seed: seed = seedClient.create(jobClient.crawlId + '_seeds', seed) return CrawlClient(self.server, seed, jobClient, rounds, index)
[ "def", "Crawl", "(", "self", ",", "seed", ",", "seedClient", "=", "None", ",", "jobClient", "=", "None", ",", "rounds", "=", "1", ",", "index", "=", "True", ")", ":", "if", "seedClient", "is", "None", ":", "seedClient", "=", "self", ".", "Seeds", "(", ")", "if", "jobClient", "is", "None", ":", "jobClient", "=", "self", ".", "Jobs", "(", ")", "if", "type", "(", "seed", ")", "!=", "Seed", ":", "seed", "=", "seedClient", ".", "create", "(", "jobClient", ".", "crawlId", "+", "'_seeds'", ",", "seed", ")", "return", "CrawlClient", "(", "self", ".", "server", ",", "seed", ",", "jobClient", ",", "rounds", ",", "index", ")" ]
Launch a crawl using the given seed :param seed: Type (Seed or SeedList) - used for crawl :param seedClient: if a SeedList is given, the SeedClient to upload, if None a default will be created :param jobClient: the JobClient to be used, if None a default will be created :param rounds: the number of rounds in the crawl :return: a CrawlClient to monitor and control the crawl
[ "Launch", "a", "crawl", "using", "the", "given", "seed", ":", "param", "seed", ":", "Type", "(", "Seed", "or", "SeedList", ")", "-", "used", "for", "crawl", ":", "param", "seedClient", ":", "if", "a", "SeedList", "is", "given", "the", "SeedClient", "to", "upload", "if", "None", "a", "default", "will", "be", "created", ":", "param", "jobClient", ":", "the", "JobClient", "to", "be", "used", "if", "None", "a", "default", "will", "be", "created", ":", "param", "rounds", ":", "the", "number", "of", "rounds", "in", "the", "crawl", ":", "return", ":", "a", "CrawlClient", "to", "monitor", "and", "control", "the", "crawl" ]
train
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L677-L693
deep-compute/logagg
logagg/formatters.py
haproxy
def haproxy(line): #TODO Handle all message formats ''' >>> import pprint >>> input_line1 = 'Apr 24 00:00:02 node haproxy[12298]: 1.1.1.1:48660 [24/Apr/2019:00:00:02.358] pre-staging~ pre-staging_doc/pre-staging_active 261/0/2/8/271 200 2406 - - ---- 4/4/0/1/0 0/0 {AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018} "GET /doc/api/get?call=apple HTTP/1.1"' >>> output_line1 = haproxy(input_line1) >>> pprint.pprint(output_line1) {'data': {'Tc': 2.0, 'Tq': 261.0, 'Tr': 8.0, 'Tw': 0.0, '_api': '/doc/api/get?call=apple', '_headers': ['AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018'], 'actconn': 4, 'backend': 'pre-staging_doc/pre-staging_active', 'backend_queue': 0, 'beconn': 1, 'bytes_read': 2406.0, 'client_port': '48660', 'client_server': '1.1.1.1', 'feconn': 4, 'front_end': 'pre-staging~', 'haproxy_server': 'node', 'method': 'GET', 'resp_time': 271.0, 'retries': 0, 'srv_conn': 0, 'srv_queue': 0, 'status': '200', 'timestamp': '2019-04-24T00:00:02.358000'}, 'event': 'haproxy_event', 'timestamp': '2019-04-24T00:00:02.358000', 'type': 'metric'} ''' _line = line.strip().split() log = {} log['client_server'] = _line[5].split(':')[0].strip() log['client_port'] = _line[5].split(':')[1].strip() _timestamp = re.findall(r'\[(.*?)\]', _line[6])[0] log['timestamp'] = datetime.datetime.strptime(_timestamp, '%d/%b/%Y:%H:%M:%S.%f').isoformat() log['front_end'] = _line[7].strip() log['backend'] = _line[8].strip() log['Tq'] = float(_line[9].split('/')[0].strip()) log['Tw'] = float(_line[9].split('/')[1].strip()) log['Tc'] = float(_line[9].split('/')[2].strip()) log['Tr'] = float(_line[9].split('/')[3].strip()) log['resp_time'] = float(_line[9].split('/')[-1].strip()) log['status'] = _line[10].strip() log['bytes_read'] = float(_line[11].strip()) log['_headers'] = re.findall(r'{(.*)}', line) log['haproxy_server'] = _line[3].strip() log['method'] = _line[-3].strip('"').strip() log['_api'] = _line[-2].strip() log['retries'] = int(_line[15].split('/')[-1].strip()) log['actconn'] = int(_line[15].split('/')[0].strip()) log['feconn'] = int(_line[15].split('/')[1].strip()) log['beconn'] = int(_line[15].split('/')[-2].strip()) log['srv_conn'] = int(_line[15].split('/')[-3].strip()) log['srv_queue'] = int(_line[16].split('/')[0].strip()) log['backend_queue'] = int(_line[16].split('/')[1].strip()) return dict( data=log, event='haproxy_event', timestamp=log.get('timestamp'), type='metric' )
python
def haproxy(line): #TODO Handle all message formats ''' >>> import pprint >>> input_line1 = 'Apr 24 00:00:02 node haproxy[12298]: 1.1.1.1:48660 [24/Apr/2019:00:00:02.358] pre-staging~ pre-staging_doc/pre-staging_active 261/0/2/8/271 200 2406 - - ---- 4/4/0/1/0 0/0 {AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018} "GET /doc/api/get?call=apple HTTP/1.1"' >>> output_line1 = haproxy(input_line1) >>> pprint.pprint(output_line1) {'data': {'Tc': 2.0, 'Tq': 261.0, 'Tr': 8.0, 'Tw': 0.0, '_api': '/doc/api/get?call=apple', '_headers': ['AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018'], 'actconn': 4, 'backend': 'pre-staging_doc/pre-staging_active', 'backend_queue': 0, 'beconn': 1, 'bytes_read': 2406.0, 'client_port': '48660', 'client_server': '1.1.1.1', 'feconn': 4, 'front_end': 'pre-staging~', 'haproxy_server': 'node', 'method': 'GET', 'resp_time': 271.0, 'retries': 0, 'srv_conn': 0, 'srv_queue': 0, 'status': '200', 'timestamp': '2019-04-24T00:00:02.358000'}, 'event': 'haproxy_event', 'timestamp': '2019-04-24T00:00:02.358000', 'type': 'metric'} ''' _line = line.strip().split() log = {} log['client_server'] = _line[5].split(':')[0].strip() log['client_port'] = _line[5].split(':')[1].strip() _timestamp = re.findall(r'\[(.*?)\]', _line[6])[0] log['timestamp'] = datetime.datetime.strptime(_timestamp, '%d/%b/%Y:%H:%M:%S.%f').isoformat() log['front_end'] = _line[7].strip() log['backend'] = _line[8].strip() log['Tq'] = float(_line[9].split('/')[0].strip()) log['Tw'] = float(_line[9].split('/')[1].strip()) log['Tc'] = float(_line[9].split('/')[2].strip()) log['Tr'] = float(_line[9].split('/')[3].strip()) log['resp_time'] = float(_line[9].split('/')[-1].strip()) log['status'] = _line[10].strip() log['bytes_read'] = float(_line[11].strip()) log['_headers'] = re.findall(r'{(.*)}', line) log['haproxy_server'] = _line[3].strip() log['method'] = _line[-3].strip('"').strip() log['_api'] = _line[-2].strip() log['retries'] = int(_line[15].split('/')[-1].strip()) log['actconn'] = int(_line[15].split('/')[0].strip()) log['feconn'] = int(_line[15].split('/')[1].strip()) log['beconn'] = int(_line[15].split('/')[-2].strip()) log['srv_conn'] = int(_line[15].split('/')[-3].strip()) log['srv_queue'] = int(_line[16].split('/')[0].strip()) log['backend_queue'] = int(_line[16].split('/')[1].strip()) return dict( data=log, event='haproxy_event', timestamp=log.get('timestamp'), type='metric' )
[ "def", "haproxy", "(", "line", ")", ":", "#TODO Handle all message formats", "_line", "=", "line", ".", "strip", "(", ")", ".", "split", "(", ")", "log", "=", "{", "}", "log", "[", "'client_server'", "]", "=", "_line", "[", "5", "]", ".", "split", "(", "':'", ")", "[", "0", "]", ".", "strip", "(", ")", "log", "[", "'client_port'", "]", "=", "_line", "[", "5", "]", ".", "split", "(", "':'", ")", "[", "1", "]", ".", "strip", "(", ")", "_timestamp", "=", "re", ".", "findall", "(", "r'\\[(.*?)\\]'", ",", "_line", "[", "6", "]", ")", "[", "0", "]", "log", "[", "'timestamp'", "]", "=", "datetime", ".", "datetime", ".", "strptime", "(", "_timestamp", ",", "'%d/%b/%Y:%H:%M:%S.%f'", ")", ".", "isoformat", "(", ")", "log", "[", "'front_end'", "]", "=", "_line", "[", "7", "]", ".", "strip", "(", ")", "log", "[", "'backend'", "]", "=", "_line", "[", "8", "]", ".", "strip", "(", ")", "log", "[", "'Tq'", "]", "=", "float", "(", "_line", "[", "9", "]", ".", "split", "(", "'/'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "log", "[", "'Tw'", "]", "=", "float", "(", "_line", "[", "9", "]", ".", "split", "(", "'/'", ")", "[", "1", "]", ".", "strip", "(", ")", ")", "log", "[", "'Tc'", "]", "=", "float", "(", "_line", "[", "9", "]", ".", "split", "(", "'/'", ")", "[", "2", "]", ".", "strip", "(", ")", ")", "log", "[", "'Tr'", "]", "=", "float", "(", "_line", "[", "9", "]", ".", "split", "(", "'/'", ")", "[", "3", "]", ".", "strip", "(", ")", ")", "log", "[", "'resp_time'", "]", "=", "float", "(", "_line", "[", "9", "]", ".", "split", "(", "'/'", ")", "[", "-", "1", "]", ".", "strip", "(", ")", ")", "log", "[", "'status'", "]", "=", "_line", "[", "10", "]", ".", "strip", "(", ")", "log", "[", "'bytes_read'", "]", "=", "float", "(", "_line", "[", "11", "]", ".", "strip", "(", ")", ")", "log", "[", "'_headers'", "]", "=", "re", ".", "findall", "(", "r'{(.*)}'", ",", "line", ")", "log", "[", "'haproxy_server'", "]", "=", "_line", "[", "3", "]", ".", "strip", "(", ")", "log", "[", "'method'", "]", "=", "_line", "[", "-", "3", "]", ".", "strip", "(", "'\"'", ")", ".", "strip", "(", ")", "log", "[", "'_api'", "]", "=", "_line", "[", "-", "2", "]", ".", "strip", "(", ")", "log", "[", "'retries'", "]", "=", "int", "(", "_line", "[", "15", "]", ".", "split", "(", "'/'", ")", "[", "-", "1", "]", ".", "strip", "(", ")", ")", "log", "[", "'actconn'", "]", "=", "int", "(", "_line", "[", "15", "]", ".", "split", "(", "'/'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "log", "[", "'feconn'", "]", "=", "int", "(", "_line", "[", "15", "]", ".", "split", "(", "'/'", ")", "[", "1", "]", ".", "strip", "(", ")", ")", "log", "[", "'beconn'", "]", "=", "int", "(", "_line", "[", "15", "]", ".", "split", "(", "'/'", ")", "[", "-", "2", "]", ".", "strip", "(", ")", ")", "log", "[", "'srv_conn'", "]", "=", "int", "(", "_line", "[", "15", "]", ".", "split", "(", "'/'", ")", "[", "-", "3", "]", ".", "strip", "(", ")", ")", "log", "[", "'srv_queue'", "]", "=", "int", "(", "_line", "[", "16", "]", ".", "split", "(", "'/'", ")", "[", "0", "]", ".", "strip", "(", ")", ")", "log", "[", "'backend_queue'", "]", "=", "int", "(", "_line", "[", "16", "]", ".", "split", "(", "'/'", ")", "[", "1", "]", ".", "strip", "(", ")", ")", "return", "dict", "(", "data", "=", "log", ",", "event", "=", "'haproxy_event'", ",", "timestamp", "=", "log", ".", "get", "(", "'timestamp'", ")", ",", "type", "=", "'metric'", ")" ]
>>> import pprint >>> input_line1 = 'Apr 24 00:00:02 node haproxy[12298]: 1.1.1.1:48660 [24/Apr/2019:00:00:02.358] pre-staging~ pre-staging_doc/pre-staging_active 261/0/2/8/271 200 2406 - - ---- 4/4/0/1/0 0/0 {AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018} "GET /doc/api/get?call=apple HTTP/1.1"' >>> output_line1 = haproxy(input_line1) >>> pprint.pprint(output_line1) {'data': {'Tc': 2.0, 'Tq': 261.0, 'Tr': 8.0, 'Tw': 0.0, '_api': '/doc/api/get?call=apple', '_headers': ['AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018'], 'actconn': 4, 'backend': 'pre-staging_doc/pre-staging_active', 'backend_queue': 0, 'beconn': 1, 'bytes_read': 2406.0, 'client_port': '48660', 'client_server': '1.1.1.1', 'feconn': 4, 'front_end': 'pre-staging~', 'haproxy_server': 'node', 'method': 'GET', 'resp_time': 271.0, 'retries': 0, 'srv_conn': 0, 'srv_queue': 0, 'status': '200', 'timestamp': '2019-04-24T00:00:02.358000'}, 'event': 'haproxy_event', 'timestamp': '2019-04-24T00:00:02.358000', 'type': 'metric'}
[ ">>>", "import", "pprint", ">>>", "input_line1", "=", "Apr", "24", "00", ":", "00", ":", "02", "node", "haproxy", "[", "12298", "]", ":", "1", ".", "1", ".", "1", ".", "1", ":", "48660", "[", "24", "/", "Apr", "/", "2019", ":", "00", ":", "00", ":", "02", ".", "358", "]", "pre", "-", "staging~", "pre", "-", "staging_doc", "/", "pre", "-", "staging_active", "261", "/", "0", "/", "2", "/", "8", "/", "271", "200", "2406", "-", "-", "----", "4", "/", "4", "/", "0", "/", "1", "/", "0", "0", "/", "0", "{", "AAAAAA", ":", "AAAAA_AAAAA", ":", "AAAAA_AAAAA_AAAAA", ":", "300A|||", "user" ]
train
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L19-L94
deep-compute/logagg
logagg/formatters.py
nginx_access
def nginx_access(line): ''' >>> import pprint >>> input_line1 = '{ \ "remote_addr": "127.0.0.1","remote_user": "-","timestamp": "1515144699.201", \ "request": "GET / HTTP/1.1","status": "200","request_time": "0.000", \ "body_bytes_sent": "396","http_referer": "-","http_user_agent": "python-requests/2.18.4", \ "http_x_forwarded_for": "-","upstream_response_time": "-" \ }' >>> output_line1 = nginx_access(input_line1) >>> pprint.pprint(output_line1) {'data': {u'body_bytes_sent': 396.0, u'http_referer': u'-', u'http_user_agent': u'python-requests/2.18.4', u'http_x_forwarded_for': u'-', u'remote_addr': u'127.0.0.1', u'remote_user': u'-', u'request': u'GET / HTTP/1.1', u'request_time': 0.0, u'status': u'200', u'timestamp': '2018-01-05T09:31:39.201000', u'upstream_response_time': 0.0}, 'event': 'nginx_event', 'timestamp': '2018-01-05T09:31:39.201000', 'type': 'metric'} >>> input_line2 = '{ \ "remote_addr": "192.158.0.51","remote_user": "-","timestamp": "1515143686.415", \ "request": "POST /mpub?topic=heartbeat HTTP/1.1","status": "404","request_time": "0.000", \ "body_bytes_sent": "152","http_referer": "-","http_user_agent": "python-requests/2.18.4", \ "http_x_forwarded_for": "-","upstream_response_time": "-" \ }' >>> output_line2 = nginx_access(input_line2) >>> pprint.pprint(output_line2) {'data': {u'body_bytes_sent': 152.0, u'http_referer': u'-', u'http_user_agent': u'python-requests/2.18.4', u'http_x_forwarded_for': u'-', u'remote_addr': u'192.158.0.51', u'remote_user': u'-', u'request': u'POST /mpub?topic=heartbeat HTTP/1.1', u'request_time': 0.0, u'status': u'404', u'timestamp': '2018-01-05T09:14:46.415000', u'upstream_response_time': 0.0}, 'event': 'nginx_event', 'timestamp': '2018-01-05T09:14:46.415000', 'type': 'metric'} ''' #TODO Handle nginx error logs log = json.loads(line) timestamp_iso = datetime.datetime.utcfromtimestamp(float(log['timestamp'])).isoformat() log.update({'timestamp':timestamp_iso}) if '-' in log.get('upstream_response_time'): log['upstream_response_time'] = 0.0 log['body_bytes_sent'] = float(log['body_bytes_sent']) log['request_time'] = float(log['request_time']) log['upstream_response_time'] = float(log['upstream_response_time']) return dict( timestamp=log.get('timestamp',' '), data=log, type='metric', event='nginx_event', )
python
def nginx_access(line): ''' >>> import pprint >>> input_line1 = '{ \ "remote_addr": "127.0.0.1","remote_user": "-","timestamp": "1515144699.201", \ "request": "GET / HTTP/1.1","status": "200","request_time": "0.000", \ "body_bytes_sent": "396","http_referer": "-","http_user_agent": "python-requests/2.18.4", \ "http_x_forwarded_for": "-","upstream_response_time": "-" \ }' >>> output_line1 = nginx_access(input_line1) >>> pprint.pprint(output_line1) {'data': {u'body_bytes_sent': 396.0, u'http_referer': u'-', u'http_user_agent': u'python-requests/2.18.4', u'http_x_forwarded_for': u'-', u'remote_addr': u'127.0.0.1', u'remote_user': u'-', u'request': u'GET / HTTP/1.1', u'request_time': 0.0, u'status': u'200', u'timestamp': '2018-01-05T09:31:39.201000', u'upstream_response_time': 0.0}, 'event': 'nginx_event', 'timestamp': '2018-01-05T09:31:39.201000', 'type': 'metric'} >>> input_line2 = '{ \ "remote_addr": "192.158.0.51","remote_user": "-","timestamp": "1515143686.415", \ "request": "POST /mpub?topic=heartbeat HTTP/1.1","status": "404","request_time": "0.000", \ "body_bytes_sent": "152","http_referer": "-","http_user_agent": "python-requests/2.18.4", \ "http_x_forwarded_for": "-","upstream_response_time": "-" \ }' >>> output_line2 = nginx_access(input_line2) >>> pprint.pprint(output_line2) {'data': {u'body_bytes_sent': 152.0, u'http_referer': u'-', u'http_user_agent': u'python-requests/2.18.4', u'http_x_forwarded_for': u'-', u'remote_addr': u'192.158.0.51', u'remote_user': u'-', u'request': u'POST /mpub?topic=heartbeat HTTP/1.1', u'request_time': 0.0, u'status': u'404', u'timestamp': '2018-01-05T09:14:46.415000', u'upstream_response_time': 0.0}, 'event': 'nginx_event', 'timestamp': '2018-01-05T09:14:46.415000', 'type': 'metric'} ''' #TODO Handle nginx error logs log = json.loads(line) timestamp_iso = datetime.datetime.utcfromtimestamp(float(log['timestamp'])).isoformat() log.update({'timestamp':timestamp_iso}) if '-' in log.get('upstream_response_time'): log['upstream_response_time'] = 0.0 log['body_bytes_sent'] = float(log['body_bytes_sent']) log['request_time'] = float(log['request_time']) log['upstream_response_time'] = float(log['upstream_response_time']) return dict( timestamp=log.get('timestamp',' '), data=log, type='metric', event='nginx_event', )
[ "def", "nginx_access", "(", "line", ")", ":", "#TODO Handle nginx error logs", "log", "=", "json", ".", "loads", "(", "line", ")", "timestamp_iso", "=", "datetime", ".", "datetime", ".", "utcfromtimestamp", "(", "float", "(", "log", "[", "'timestamp'", "]", ")", ")", ".", "isoformat", "(", ")", "log", ".", "update", "(", "{", "'timestamp'", ":", "timestamp_iso", "}", ")", "if", "'-'", "in", "log", ".", "get", "(", "'upstream_response_time'", ")", ":", "log", "[", "'upstream_response_time'", "]", "=", "0.0", "log", "[", "'body_bytes_sent'", "]", "=", "float", "(", "log", "[", "'body_bytes_sent'", "]", ")", "log", "[", "'request_time'", "]", "=", "float", "(", "log", "[", "'request_time'", "]", ")", "log", "[", "'upstream_response_time'", "]", "=", "float", "(", "log", "[", "'upstream_response_time'", "]", ")", "return", "dict", "(", "timestamp", "=", "log", ".", "get", "(", "'timestamp'", ",", "' '", ")", ",", "data", "=", "log", ",", "type", "=", "'metric'", ",", "event", "=", "'nginx_event'", ",", ")" ]
>>> import pprint >>> input_line1 = '{ \ "remote_addr": "127.0.0.1","remote_user": "-","timestamp": "1515144699.201", \ "request": "GET / HTTP/1.1","status": "200","request_time": "0.000", \ "body_bytes_sent": "396","http_referer": "-","http_user_agent": "python-requests/2.18.4", \ "http_x_forwarded_for": "-","upstream_response_time": "-" \ }' >>> output_line1 = nginx_access(input_line1) >>> pprint.pprint(output_line1) {'data': {u'body_bytes_sent': 396.0, u'http_referer': u'-', u'http_user_agent': u'python-requests/2.18.4', u'http_x_forwarded_for': u'-', u'remote_addr': u'127.0.0.1', u'remote_user': u'-', u'request': u'GET / HTTP/1.1', u'request_time': 0.0, u'status': u'200', u'timestamp': '2018-01-05T09:31:39.201000', u'upstream_response_time': 0.0}, 'event': 'nginx_event', 'timestamp': '2018-01-05T09:31:39.201000', 'type': 'metric'} >>> input_line2 = '{ \ "remote_addr": "192.158.0.51","remote_user": "-","timestamp": "1515143686.415", \ "request": "POST /mpub?topic=heartbeat HTTP/1.1","status": "404","request_time": "0.000", \ "body_bytes_sent": "152","http_referer": "-","http_user_agent": "python-requests/2.18.4", \ "http_x_forwarded_for": "-","upstream_response_time": "-" \ }' >>> output_line2 = nginx_access(input_line2) >>> pprint.pprint(output_line2) {'data': {u'body_bytes_sent': 152.0, u'http_referer': u'-', u'http_user_agent': u'python-requests/2.18.4', u'http_x_forwarded_for': u'-', u'remote_addr': u'192.158.0.51', u'remote_user': u'-', u'request': u'POST /mpub?topic=heartbeat HTTP/1.1', u'request_time': 0.0, u'status': u'404', u'timestamp': '2018-01-05T09:14:46.415000', u'upstream_response_time': 0.0}, 'event': 'nginx_event', 'timestamp': '2018-01-05T09:14:46.415000', 'type': 'metric'}
[ ">>>", "import", "pprint", ">>>", "input_line1", "=", "{", "\\", "remote_addr", ":", "127", ".", "0", ".", "0", ".", "1", "remote_user", ":", "-", "timestamp", ":", "1515144699", ".", "201", "\\", "request", ":", "GET", "/", "HTTP", "/", "1", ".", "1", "status", ":", "200", "request_time", ":", "0", ".", "000", "\\", "body_bytes_sent", ":", "396", "http_referer", ":", "-", "http_user_agent", ":", "python", "-", "requests", "/", "2", ".", "18", ".", "4", "\\", "http_x_forwarded_for", ":", "-", "upstream_response_time", ":", "-", "\\", "}", ">>>", "output_line1", "=", "nginx_access", "(", "input_line1", ")", ">>>", "pprint", ".", "pprint", "(", "output_line1", ")", "{", "data", ":", "{", "u", "body_bytes_sent", ":", "396", ".", "0", "u", "http_referer", ":", "u", "-", "u", "http_user_agent", ":", "u", "python", "-", "requests", "/", "2", ".", "18", ".", "4", "u", "http_x_forwarded_for", ":", "u", "-", "u", "remote_addr", ":", "u", "127", ".", "0", ".", "0", ".", "1", "u", "remote_user", ":", "u", "-", "u", "request", ":", "u", "GET", "/", "HTTP", "/", "1", ".", "1", "u", "request_time", ":", "0", ".", "0", "u", "status", ":", "u", "200", "u", "timestamp", ":", "2018", "-", "01", "-", "05T09", ":", "31", ":", "39", ".", "201000", "u", "upstream_response_time", ":", "0", ".", "0", "}", "event", ":", "nginx_event", "timestamp", ":", "2018", "-", "01", "-", "05T09", ":", "31", ":", "39", ".", "201000", "type", ":", "metric", "}" ]
train
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L96-L160
deep-compute/logagg
logagg/formatters.py
mongodb
def mongodb(line): ''' >>> import pprint >>> input_line1 = '2017-08-17T07:56:33.489+0200 I REPL [signalProcessingThread] shutting down replication subsystems' >>> output_line1 = mongodb(input_line1) >>> pprint.pprint(output_line1) {'data': {'component': 'REPL', 'context': '[signalProcessingThread]', 'message': 'shutting down replication subsystems', 'severity': 'I', 'timestamp': '2017-08-17T07:56:33.489+0200'}, 'timestamp': '2017-08-17T07:56:33.489+0200', 'type': 'log'} >>> input_line2 = '2017-08-17T07:56:33.515+0200 W NETWORK [initandlisten] No primary detected for set confsvr_repl1' >>> output_line2 = mongodb(input_line2) >>> pprint.pprint(output_line2) {'data': {'component': 'NETWORK', 'context': '[initandlisten]', 'message': 'No primary detected for set confsvr_repl1', 'severity': 'W', 'timestamp': '2017-08-17T07:56:33.515+0200'}, 'timestamp': '2017-08-17T07:56:33.515+0200', 'type': 'log'} ''' keys = ['timestamp', 'severity', 'component', 'context', 'message'] values = re.split(r'\s+', line, maxsplit=4) mongodb_log = dict(zip(keys,values)) return dict( timestamp=values[0], data=mongodb_log, type='log', )
python
def mongodb(line): ''' >>> import pprint >>> input_line1 = '2017-08-17T07:56:33.489+0200 I REPL [signalProcessingThread] shutting down replication subsystems' >>> output_line1 = mongodb(input_line1) >>> pprint.pprint(output_line1) {'data': {'component': 'REPL', 'context': '[signalProcessingThread]', 'message': 'shutting down replication subsystems', 'severity': 'I', 'timestamp': '2017-08-17T07:56:33.489+0200'}, 'timestamp': '2017-08-17T07:56:33.489+0200', 'type': 'log'} >>> input_line2 = '2017-08-17T07:56:33.515+0200 W NETWORK [initandlisten] No primary detected for set confsvr_repl1' >>> output_line2 = mongodb(input_line2) >>> pprint.pprint(output_line2) {'data': {'component': 'NETWORK', 'context': '[initandlisten]', 'message': 'No primary detected for set confsvr_repl1', 'severity': 'W', 'timestamp': '2017-08-17T07:56:33.515+0200'}, 'timestamp': '2017-08-17T07:56:33.515+0200', 'type': 'log'} ''' keys = ['timestamp', 'severity', 'component', 'context', 'message'] values = re.split(r'\s+', line, maxsplit=4) mongodb_log = dict(zip(keys,values)) return dict( timestamp=values[0], data=mongodb_log, type='log', )
[ "def", "mongodb", "(", "line", ")", ":", "keys", "=", "[", "'timestamp'", ",", "'severity'", ",", "'component'", ",", "'context'", ",", "'message'", "]", "values", "=", "re", ".", "split", "(", "r'\\s+'", ",", "line", ",", "maxsplit", "=", "4", ")", "mongodb_log", "=", "dict", "(", "zip", "(", "keys", ",", "values", ")", ")", "return", "dict", "(", "timestamp", "=", "values", "[", "0", "]", ",", "data", "=", "mongodb_log", ",", "type", "=", "'log'", ",", ")" ]
>>> import pprint >>> input_line1 = '2017-08-17T07:56:33.489+0200 I REPL [signalProcessingThread] shutting down replication subsystems' >>> output_line1 = mongodb(input_line1) >>> pprint.pprint(output_line1) {'data': {'component': 'REPL', 'context': '[signalProcessingThread]', 'message': 'shutting down replication subsystems', 'severity': 'I', 'timestamp': '2017-08-17T07:56:33.489+0200'}, 'timestamp': '2017-08-17T07:56:33.489+0200', 'type': 'log'} >>> input_line2 = '2017-08-17T07:56:33.515+0200 W NETWORK [initandlisten] No primary detected for set confsvr_repl1' >>> output_line2 = mongodb(input_line2) >>> pprint.pprint(output_line2) {'data': {'component': 'NETWORK', 'context': '[initandlisten]', 'message': 'No primary detected for set confsvr_repl1', 'severity': 'W', 'timestamp': '2017-08-17T07:56:33.515+0200'}, 'timestamp': '2017-08-17T07:56:33.515+0200', 'type': 'log'}
[ ">>>", "import", "pprint", ">>>", "input_line1", "=", "2017", "-", "08", "-", "17T07", ":", "56", ":", "33", ".", "489", "+", "0200", "I", "REPL", "[", "signalProcessingThread", "]", "shutting", "down", "replication", "subsystems", ">>>", "output_line1", "=", "mongodb", "(", "input_line1", ")", ">>>", "pprint", ".", "pprint", "(", "output_line1", ")", "{", "data", ":", "{", "component", ":", "REPL", "context", ":", "[", "signalProcessingThread", "]", "message", ":", "shutting", "down", "replication", "subsystems", "severity", ":", "I", "timestamp", ":", "2017", "-", "08", "-", "17T07", ":", "56", ":", "33", ".", "489", "+", "0200", "}", "timestamp", ":", "2017", "-", "08", "-", "17T07", ":", "56", ":", "33", ".", "489", "+", "0200", "type", ":", "log", "}" ]
train
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L162-L196
deep-compute/logagg
logagg/formatters.py
django
def django(line): ''' >>> import pprint >>> input_line1 = '[23/Aug/2017 11:35:25] INFO [app.middleware_log_req:50]View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }' >>> output_line1 = django(input_line1) >>> pprint.pprint(output_line1) {'data': {'loglevel': 'INFO', 'logname': '[app.middleware_log_req:50]', 'message': 'View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }', 'timestamp': '2017-08-23T11:35:25'}, 'level': 'INFO', 'timestamp': '2017-08-23T11:35:25'} >>> input_line2 = '[22/Sep/2017 06:32:15] INFO [app.function:6022] {"UUID": "c47f3530-9f5f-11e7-a559-917d011459f7", "timestamp":1506061932546, "misc": {"status": 200, "ready_state": 4, "end_time_ms": 1506061932546, "url": "/api/function?", "start_time_ms": 1506061932113, "response_length": 31, "status_message": "OK", "request_time_ms": 433}, "user": "root", "host_url": "localhost:8888", "message": "ajax success"}' >>> output_line2 = django(input_line2) >>> pprint.pprint(output_line2) {'data': {'loglevel': 'INFO', 'logname': '[app.function:6022]', 'message': {u'UUID': u'c47f3530-9f5f-11e7-a559-917d011459f7', u'host_url': u'localhost:8888', u'message': u'ajax success', u'misc': {u'end_time_ms': 1506061932546L, u'ready_state': 4, u'request_time_ms': 433, u'response_length': 31, u'start_time_ms': 1506061932113L, u'status': 200, u'status_message': u'OK', u'url': u'/api/function?'}, u'timestamp': 1506061932546L, u'user': u'root'}, 'timestamp': '2017-09-22T06:32:15'}, 'level': 'INFO', 'timestamp': '2017-09-22T06:32:15'} Case2: [18/Sep/2017 05:40:36] ERROR [app.apps:78] failed to get the record, collection = Collection(Database(MongoClient(host=['localhost:27017'], document_class=dict, tz_aware=False, connect=True, serverselectiontimeoutms=3000), u'collection_cache'), u'function_dummy_version') Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/mongo_cache/mongocache.py", line 70, in __getitem__ result = self.collection.find_one({"_id": key}) OperationFailure: not authorized on collection_cache to execute command { find: "function", filter: { _id: "zydelig-cosine-20" }, limit: 1, singleBatch: true } ''' #TODO we need to handle case2 logs data = {} log = re.findall(r'^(\[\d+/\w+/\d+ \d+:\d+:\d+\].*)', line) if len(log) == 1: data['timestamp'] = datetime.datetime.strptime(re.findall(r'(\d+/\w+/\d+ \d+:\d+:\d+)',\ log[0])[0],"%d/%b/%Y %H:%M:%S").isoformat() data['loglevel'] = re.findall('[A-Z]+', log[0])[1] data['logname'] = re.findall('\[\D+.\w+:\d+\]', log[0])[0] message = re.findall('\{.+\}', log[0]) try: if len(message) > 0: message = json.loads(message[0]) else: message = re.split(']', log[0]) message = ''.join(message[2:]) except ValueError: message = re.split(']', log[0]) message = ''.join(message[2:]) data['message'] = message return dict( timestamp=data['timestamp'], level=data['loglevel'], data=data, ) else: return dict( timestamp=datetime.datetime.isoformat(datetime.datetime.utcnow()), data={raw:line} )
python
def django(line): ''' >>> import pprint >>> input_line1 = '[23/Aug/2017 11:35:25] INFO [app.middleware_log_req:50]View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }' >>> output_line1 = django(input_line1) >>> pprint.pprint(output_line1) {'data': {'loglevel': 'INFO', 'logname': '[app.middleware_log_req:50]', 'message': 'View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }', 'timestamp': '2017-08-23T11:35:25'}, 'level': 'INFO', 'timestamp': '2017-08-23T11:35:25'} >>> input_line2 = '[22/Sep/2017 06:32:15] INFO [app.function:6022] {"UUID": "c47f3530-9f5f-11e7-a559-917d011459f7", "timestamp":1506061932546, "misc": {"status": 200, "ready_state": 4, "end_time_ms": 1506061932546, "url": "/api/function?", "start_time_ms": 1506061932113, "response_length": 31, "status_message": "OK", "request_time_ms": 433}, "user": "root", "host_url": "localhost:8888", "message": "ajax success"}' >>> output_line2 = django(input_line2) >>> pprint.pprint(output_line2) {'data': {'loglevel': 'INFO', 'logname': '[app.function:6022]', 'message': {u'UUID': u'c47f3530-9f5f-11e7-a559-917d011459f7', u'host_url': u'localhost:8888', u'message': u'ajax success', u'misc': {u'end_time_ms': 1506061932546L, u'ready_state': 4, u'request_time_ms': 433, u'response_length': 31, u'start_time_ms': 1506061932113L, u'status': 200, u'status_message': u'OK', u'url': u'/api/function?'}, u'timestamp': 1506061932546L, u'user': u'root'}, 'timestamp': '2017-09-22T06:32:15'}, 'level': 'INFO', 'timestamp': '2017-09-22T06:32:15'} Case2: [18/Sep/2017 05:40:36] ERROR [app.apps:78] failed to get the record, collection = Collection(Database(MongoClient(host=['localhost:27017'], document_class=dict, tz_aware=False, connect=True, serverselectiontimeoutms=3000), u'collection_cache'), u'function_dummy_version') Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/mongo_cache/mongocache.py", line 70, in __getitem__ result = self.collection.find_one({"_id": key}) OperationFailure: not authorized on collection_cache to execute command { find: "function", filter: { _id: "zydelig-cosine-20" }, limit: 1, singleBatch: true } ''' #TODO we need to handle case2 logs data = {} log = re.findall(r'^(\[\d+/\w+/\d+ \d+:\d+:\d+\].*)', line) if len(log) == 1: data['timestamp'] = datetime.datetime.strptime(re.findall(r'(\d+/\w+/\d+ \d+:\d+:\d+)',\ log[0])[0],"%d/%b/%Y %H:%M:%S").isoformat() data['loglevel'] = re.findall('[A-Z]+', log[0])[1] data['logname'] = re.findall('\[\D+.\w+:\d+\]', log[0])[0] message = re.findall('\{.+\}', log[0]) try: if len(message) > 0: message = json.loads(message[0]) else: message = re.split(']', log[0]) message = ''.join(message[2:]) except ValueError: message = re.split(']', log[0]) message = ''.join(message[2:]) data['message'] = message return dict( timestamp=data['timestamp'], level=data['loglevel'], data=data, ) else: return dict( timestamp=datetime.datetime.isoformat(datetime.datetime.utcnow()), data={raw:line} )
[ "def", "django", "(", "line", ")", ":", "#TODO we need to handle case2 logs", "data", "=", "{", "}", "log", "=", "re", ".", "findall", "(", "r'^(\\[\\d+/\\w+/\\d+ \\d+:\\d+:\\d+\\].*)'", ",", "line", ")", "if", "len", "(", "log", ")", "==", "1", ":", "data", "[", "'timestamp'", "]", "=", "datetime", ".", "datetime", ".", "strptime", "(", "re", ".", "findall", "(", "r'(\\d+/\\w+/\\d+ \\d+:\\d+:\\d+)'", ",", "log", "[", "0", "]", ")", "[", "0", "]", ",", "\"%d/%b/%Y %H:%M:%S\"", ")", ".", "isoformat", "(", ")", "data", "[", "'loglevel'", "]", "=", "re", ".", "findall", "(", "'[A-Z]+'", ",", "log", "[", "0", "]", ")", "[", "1", "]", "data", "[", "'logname'", "]", "=", "re", ".", "findall", "(", "'\\[\\D+.\\w+:\\d+\\]'", ",", "log", "[", "0", "]", ")", "[", "0", "]", "message", "=", "re", ".", "findall", "(", "'\\{.+\\}'", ",", "log", "[", "0", "]", ")", "try", ":", "if", "len", "(", "message", ")", ">", "0", ":", "message", "=", "json", ".", "loads", "(", "message", "[", "0", "]", ")", "else", ":", "message", "=", "re", ".", "split", "(", "']'", ",", "log", "[", "0", "]", ")", "message", "=", "''", ".", "join", "(", "message", "[", "2", ":", "]", ")", "except", "ValueError", ":", "message", "=", "re", ".", "split", "(", "']'", ",", "log", "[", "0", "]", ")", "message", "=", "''", ".", "join", "(", "message", "[", "2", ":", "]", ")", "data", "[", "'message'", "]", "=", "message", "return", "dict", "(", "timestamp", "=", "data", "[", "'timestamp'", "]", ",", "level", "=", "data", "[", "'loglevel'", "]", ",", "data", "=", "data", ",", ")", "else", ":", "return", "dict", "(", "timestamp", "=", "datetime", ".", "datetime", ".", "isoformat", "(", "datetime", ".", "datetime", ".", "utcnow", "(", ")", ")", ",", "data", "=", "{", "raw", ":", "line", "}", ")" ]
>>> import pprint >>> input_line1 = '[23/Aug/2017 11:35:25] INFO [app.middleware_log_req:50]View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }' >>> output_line1 = django(input_line1) >>> pprint.pprint(output_line1) {'data': {'loglevel': 'INFO', 'logname': '[app.middleware_log_req:50]', 'message': 'View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }', 'timestamp': '2017-08-23T11:35:25'}, 'level': 'INFO', 'timestamp': '2017-08-23T11:35:25'} >>> input_line2 = '[22/Sep/2017 06:32:15] INFO [app.function:6022] {"UUID": "c47f3530-9f5f-11e7-a559-917d011459f7", "timestamp":1506061932546, "misc": {"status": 200, "ready_state": 4, "end_time_ms": 1506061932546, "url": "/api/function?", "start_time_ms": 1506061932113, "response_length": 31, "status_message": "OK", "request_time_ms": 433}, "user": "root", "host_url": "localhost:8888", "message": "ajax success"}' >>> output_line2 = django(input_line2) >>> pprint.pprint(output_line2) {'data': {'loglevel': 'INFO', 'logname': '[app.function:6022]', 'message': {u'UUID': u'c47f3530-9f5f-11e7-a559-917d011459f7', u'host_url': u'localhost:8888', u'message': u'ajax success', u'misc': {u'end_time_ms': 1506061932546L, u'ready_state': 4, u'request_time_ms': 433, u'response_length': 31, u'start_time_ms': 1506061932113L, u'status': 200, u'status_message': u'OK', u'url': u'/api/function?'}, u'timestamp': 1506061932546L, u'user': u'root'}, 'timestamp': '2017-09-22T06:32:15'}, 'level': 'INFO', 'timestamp': '2017-09-22T06:32:15'} Case2: [18/Sep/2017 05:40:36] ERROR [app.apps:78] failed to get the record, collection = Collection(Database(MongoClient(host=['localhost:27017'], document_class=dict, tz_aware=False, connect=True, serverselectiontimeoutms=3000), u'collection_cache'), u'function_dummy_version') Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/mongo_cache/mongocache.py", line 70, in __getitem__ result = self.collection.find_one({"_id": key}) OperationFailure: not authorized on collection_cache to execute command { find: "function", filter: { _id: "zydelig-cosine-20" }, limit: 1, singleBatch: true }
[ ">>>", "import", "pprint", ">>>", "input_line1", "=", "[", "23", "/", "Aug", "/", "2017", "11", ":", "35", ":", "25", "]", "INFO", "[", "app", ".", "middleware_log_req", ":", "50", "]", "View", "func", "called", ":", "{", "exception", ":", "null", "processing_time", ":", "0", ".", "00011801719665527344", "url", ":", "<url", ">", "host", ":", "localhost", "user", ":", "testing", "post_contents", ":", "method", ":", "POST", "}", ">>>", "output_line1", "=", "django", "(", "input_line1", ")", ">>>", "pprint", ".", "pprint", "(", "output_line1", ")", "{", "data", ":", "{", "loglevel", ":", "INFO", "logname", ":", "[", "app", ".", "middleware_log_req", ":", "50", "]", "message", ":", "View", "func", "called", ":", "{", "exception", ":", "null", "processing_time", ":", "0", ".", "00011801719665527344", "url", ":", "<url", ">", "host", ":", "localhost", "user", ":", "testing", "post_contents", ":", "method", ":", "POST", "}", "timestamp", ":", "2017", "-", "08", "-", "23T11", ":", "35", ":", "25", "}", "level", ":", "INFO", "timestamp", ":", "2017", "-", "08", "-", "23T11", ":", "35", ":", "25", "}" ]
train
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L199-L271
deep-compute/logagg
logagg/formatters.py
basescript
def basescript(line): ''' >>> import pprint >>> input_line = '{"level": "warning", "timestamp": "2018-02-07T06:37:00.297610Z", "event": "exited via keyboard interrupt", "type": "log", "id": "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_": {"ln": 58, "file": "/usr/local/lib/python2.7/dist-packages/basescript/basescript.py", "name": "basescript.basescript", "fn": "start"}}' >>> output_line1 = basescript(input_line) >>> pprint.pprint(output_line1) {'data': {u'_': {u'file': u'/usr/local/lib/python2.7/dist-packages/basescript/basescript.py', u'fn': u'start', u'ln': 58, u'name': u'basescript.basescript'}, u'event': u'exited via keyboard interrupt', u'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65', u'level': u'warning', u'timestamp': u'2018-02-07T06:37:00.297610Z', u'type': u'log'}, 'event': u'exited via keyboard interrupt', 'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65', 'level': u'warning', 'timestamp': u'2018-02-07T06:37:00.297610Z', 'type': u'log'} ''' log = json.loads(line) return dict( timestamp=log['timestamp'], data=log, id=log['id'], type=log['type'], level=log['level'], event=log['event'] )
python
def basescript(line): ''' >>> import pprint >>> input_line = '{"level": "warning", "timestamp": "2018-02-07T06:37:00.297610Z", "event": "exited via keyboard interrupt", "type": "log", "id": "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_": {"ln": 58, "file": "/usr/local/lib/python2.7/dist-packages/basescript/basescript.py", "name": "basescript.basescript", "fn": "start"}}' >>> output_line1 = basescript(input_line) >>> pprint.pprint(output_line1) {'data': {u'_': {u'file': u'/usr/local/lib/python2.7/dist-packages/basescript/basescript.py', u'fn': u'start', u'ln': 58, u'name': u'basescript.basescript'}, u'event': u'exited via keyboard interrupt', u'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65', u'level': u'warning', u'timestamp': u'2018-02-07T06:37:00.297610Z', u'type': u'log'}, 'event': u'exited via keyboard interrupt', 'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65', 'level': u'warning', 'timestamp': u'2018-02-07T06:37:00.297610Z', 'type': u'log'} ''' log = json.loads(line) return dict( timestamp=log['timestamp'], data=log, id=log['id'], type=log['type'], level=log['level'], event=log['event'] )
[ "def", "basescript", "(", "line", ")", ":", "log", "=", "json", ".", "loads", "(", "line", ")", "return", "dict", "(", "timestamp", "=", "log", "[", "'timestamp'", "]", ",", "data", "=", "log", ",", "id", "=", "log", "[", "'id'", "]", ",", "type", "=", "log", "[", "'type'", "]", ",", "level", "=", "log", "[", "'level'", "]", ",", "event", "=", "log", "[", "'event'", "]", ")" ]
>>> import pprint >>> input_line = '{"level": "warning", "timestamp": "2018-02-07T06:37:00.297610Z", "event": "exited via keyboard interrupt", "type": "log", "id": "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_": {"ln": 58, "file": "/usr/local/lib/python2.7/dist-packages/basescript/basescript.py", "name": "basescript.basescript", "fn": "start"}}' >>> output_line1 = basescript(input_line) >>> pprint.pprint(output_line1) {'data': {u'_': {u'file': u'/usr/local/lib/python2.7/dist-packages/basescript/basescript.py', u'fn': u'start', u'ln': 58, u'name': u'basescript.basescript'}, u'event': u'exited via keyboard interrupt', u'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65', u'level': u'warning', u'timestamp': u'2018-02-07T06:37:00.297610Z', u'type': u'log'}, 'event': u'exited via keyboard interrupt', 'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65', 'level': u'warning', 'timestamp': u'2018-02-07T06:37:00.297610Z', 'type': u'log'}
[ ">>>", "import", "pprint", ">>>", "input_line", "=", "{", "level", ":", "warning", "timestamp", ":", "2018", "-", "02", "-", "07T06", ":", "37", ":", "00", ".", "297610Z", "event", ":", "exited", "via", "keyboard", "interrupt", "type", ":", "log", "id", ":", "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_", ":", "{", "ln", ":", "58", "file", ":", "/", "usr", "/", "local", "/", "lib", "/", "python2", ".", "7", "/", "dist", "-", "packages", "/", "basescript", "/", "basescript", ".", "py", "name", ":", "basescript", ".", "basescript", "fn", ":", "start", "}}", ">>>", "output_line1", "=", "basescript", "(", "input_line", ")", ">>>", "pprint", ".", "pprint", "(", "output_line1", ")", "{", "data", ":", "{", "u", "_", ":", "{", "u", "file", ":", "u", "/", "usr", "/", "local", "/", "lib", "/", "python2", ".", "7", "/", "dist", "-", "packages", "/", "basescript", "/", "basescript", ".", "py", "u", "fn", ":", "u", "start", "u", "ln", ":", "58", "u", "name", ":", "u", "basescript", ".", "basescript", "}", "u", "event", ":", "u", "exited", "via", "keyboard", "interrupt", "u", "id", ":", "u", "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "u", "level", ":", "u", "warning", "u", "timestamp", ":", "u", "2018", "-", "02", "-", "07T06", ":", "37", ":", "00", ".", "297610Z", "u", "type", ":", "u", "log", "}", "event", ":", "u", "exited", "via", "keyboard", "interrupt", "id", ":", "u", "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "level", ":", "u", "warning", "timestamp", ":", "u", "2018", "-", "02", "-", "07T06", ":", "37", ":", "00", ".", "297610Z", "type", ":", "u", "log", "}" ]
train
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L273-L304