repository_name
stringlengths 7
55
| func_path_in_repository
stringlengths 4
223
| func_name
stringlengths 1
134
| whole_func_string
stringlengths 75
104k
| language
stringclasses 1
value | func_code_string
stringlengths 75
104k
| func_code_tokens
listlengths 19
28.4k
| func_documentation_string
stringlengths 1
46.9k
| func_documentation_tokens
listlengths 1
1.97k
| split_name
stringclasses 1
value | func_code_url
stringlengths 87
315
|
|---|---|---|---|---|---|---|---|---|---|---|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.start_engine
|
def start_engine(self, **kwargs):
"""
Initializes the workflow with given request, response objects and diagram name.
Args:
session:
input:
workflow_name (str): Name of workflow diagram without ".bpmn" suffix.
File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
"""
self.current = WFCurrent(**kwargs)
self.wf_state = {'in_external': False, 'finished': False}
if not self.current.new_token:
self.wf_state = self.current.wf_cache.get(self.wf_state)
self.current.workflow_name = self.wf_state['name']
# if we have a pre-selected object to work with,
# inserting it as current.input['id'] and task_data['object_id']
if 'subject' in self.wf_state:
self.current.input['id'] = self.wf_state['subject']
self.current.task_data['object_id'] = self.wf_state['subject']
self.check_for_authentication()
self.check_for_permission()
self.workflow = self.load_or_create_workflow()
# if form data exists in input (user submitted)
# put form data in wf task_data
if 'form' in self.current.input:
form = self.current.input['form']
if 'form_name' in form:
self.current.task_data[form['form_name']] = form
# in wf diagram, if property is stated as init = True
# demanded initial values are assigned and put to cache
start_init_values = self.workflow_spec.wf_properties.get('init', 'False') == 'True'
if start_init_values:
WFInit = get_object_from_path(settings.WF_INITIAL_VALUES)()
WFInit.assign_wf_initial_values(self.current)
log_msg = ("\n\n::::::::::: ENGINE STARTED :::::::::::\n"
"\tWF: %s (Possible) TASK:%s\n"
"\tCMD:%s\n"
"\tSUBCMD:%s" % (
self.workflow.name,
self.workflow.get_tasks(Task.READY),
self.current.input.get('cmd'), self.current.input.get('subcmd')))
log.debug(log_msg)
sys._zops_wf_state_log = log_msg
self.current.workflow = self.workflow
|
python
|
def start_engine(self, **kwargs):
"""
Initializes the workflow with given request, response objects and diagram name.
Args:
session:
input:
workflow_name (str): Name of workflow diagram without ".bpmn" suffix.
File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
"""
self.current = WFCurrent(**kwargs)
self.wf_state = {'in_external': False, 'finished': False}
if not self.current.new_token:
self.wf_state = self.current.wf_cache.get(self.wf_state)
self.current.workflow_name = self.wf_state['name']
# if we have a pre-selected object to work with,
# inserting it as current.input['id'] and task_data['object_id']
if 'subject' in self.wf_state:
self.current.input['id'] = self.wf_state['subject']
self.current.task_data['object_id'] = self.wf_state['subject']
self.check_for_authentication()
self.check_for_permission()
self.workflow = self.load_or_create_workflow()
# if form data exists in input (user submitted)
# put form data in wf task_data
if 'form' in self.current.input:
form = self.current.input['form']
if 'form_name' in form:
self.current.task_data[form['form_name']] = form
# in wf diagram, if property is stated as init = True
# demanded initial values are assigned and put to cache
start_init_values = self.workflow_spec.wf_properties.get('init', 'False') == 'True'
if start_init_values:
WFInit = get_object_from_path(settings.WF_INITIAL_VALUES)()
WFInit.assign_wf_initial_values(self.current)
log_msg = ("\n\n::::::::::: ENGINE STARTED :::::::::::\n"
"\tWF: %s (Possible) TASK:%s\n"
"\tCMD:%s\n"
"\tSUBCMD:%s" % (
self.workflow.name,
self.workflow.get_tasks(Task.READY),
self.current.input.get('cmd'), self.current.input.get('subcmd')))
log.debug(log_msg)
sys._zops_wf_state_log = log_msg
self.current.workflow = self.workflow
|
[
"def",
"start_engine",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"self",
".",
"current",
"=",
"WFCurrent",
"(",
"*",
"*",
"kwargs",
")",
"self",
".",
"wf_state",
"=",
"{",
"'in_external'",
":",
"False",
",",
"'finished'",
":",
"False",
"}",
"if",
"not",
"self",
".",
"current",
".",
"new_token",
":",
"self",
".",
"wf_state",
"=",
"self",
".",
"current",
".",
"wf_cache",
".",
"get",
"(",
"self",
".",
"wf_state",
")",
"self",
".",
"current",
".",
"workflow_name",
"=",
"self",
".",
"wf_state",
"[",
"'name'",
"]",
"# if we have a pre-selected object to work with,",
"# inserting it as current.input['id'] and task_data['object_id']",
"if",
"'subject'",
"in",
"self",
".",
"wf_state",
":",
"self",
".",
"current",
".",
"input",
"[",
"'id'",
"]",
"=",
"self",
".",
"wf_state",
"[",
"'subject'",
"]",
"self",
".",
"current",
".",
"task_data",
"[",
"'object_id'",
"]",
"=",
"self",
".",
"wf_state",
"[",
"'subject'",
"]",
"self",
".",
"check_for_authentication",
"(",
")",
"self",
".",
"check_for_permission",
"(",
")",
"self",
".",
"workflow",
"=",
"self",
".",
"load_or_create_workflow",
"(",
")",
"# if form data exists in input (user submitted)",
"# put form data in wf task_data",
"if",
"'form'",
"in",
"self",
".",
"current",
".",
"input",
":",
"form",
"=",
"self",
".",
"current",
".",
"input",
"[",
"'form'",
"]",
"if",
"'form_name'",
"in",
"form",
":",
"self",
".",
"current",
".",
"task_data",
"[",
"form",
"[",
"'form_name'",
"]",
"]",
"=",
"form",
"# in wf diagram, if property is stated as init = True",
"# demanded initial values are assigned and put to cache",
"start_init_values",
"=",
"self",
".",
"workflow_spec",
".",
"wf_properties",
".",
"get",
"(",
"'init'",
",",
"'False'",
")",
"==",
"'True'",
"if",
"start_init_values",
":",
"WFInit",
"=",
"get_object_from_path",
"(",
"settings",
".",
"WF_INITIAL_VALUES",
")",
"(",
")",
"WFInit",
".",
"assign_wf_initial_values",
"(",
"self",
".",
"current",
")",
"log_msg",
"=",
"(",
"\"\\n\\n::::::::::: ENGINE STARTED :::::::::::\\n\"",
"\"\\tWF: %s (Possible) TASK:%s\\n\"",
"\"\\tCMD:%s\\n\"",
"\"\\tSUBCMD:%s\"",
"%",
"(",
"self",
".",
"workflow",
".",
"name",
",",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"Task",
".",
"READY",
")",
",",
"self",
".",
"current",
".",
"input",
".",
"get",
"(",
"'cmd'",
")",
",",
"self",
".",
"current",
".",
"input",
".",
"get",
"(",
"'subcmd'",
")",
")",
")",
"log",
".",
"debug",
"(",
"log_msg",
")",
"sys",
".",
"_zops_wf_state_log",
"=",
"log_msg",
"self",
".",
"current",
".",
"workflow",
"=",
"self",
".",
"workflow"
] |
Initializes the workflow with given request, response objects and diagram name.
Args:
session:
input:
workflow_name (str): Name of workflow diagram without ".bpmn" suffix.
File must be placed under one of configured :py:attr:`~zengine.settings.WORKFLOW_PACKAGES_PATHS`
|
[
"Initializes",
"the",
"workflow",
"with",
"given",
"request",
"response",
"objects",
"and",
"diagram",
"name",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L241-L289
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.generate_wf_state_log
|
def generate_wf_state_log(self):
"""
Logs the state of workflow and content of task_data.
"""
output = '\n- - - - - -\n'
output += "WORKFLOW: %s ( %s )" % (self.current.workflow_name.upper(),
self.current.workflow.name)
output += "\nTASK: %s ( %s )\n" % (self.current.task_name, self.current.task_type)
output += "DATA:"
for k, v in self.current.task_data.items():
if v:
output += "\n\t%s: %s" % (k, v)
output += "\nCURRENT:"
output += "\n\tACTIVITY: %s" % self.current.activity
output += "\n\tPOOL: %s" % self.current.pool
output += "\n\tIN EXTERNAL: %s" % self.wf_state['in_external']
output += "\n\tLANE: %s" % self.current.lane_name
output += "\n\tTOKEN: %s" % self.current.token
sys._zops_wf_state_log = output
return output
|
python
|
def generate_wf_state_log(self):
"""
Logs the state of workflow and content of task_data.
"""
output = '\n- - - - - -\n'
output += "WORKFLOW: %s ( %s )" % (self.current.workflow_name.upper(),
self.current.workflow.name)
output += "\nTASK: %s ( %s )\n" % (self.current.task_name, self.current.task_type)
output += "DATA:"
for k, v in self.current.task_data.items():
if v:
output += "\n\t%s: %s" % (k, v)
output += "\nCURRENT:"
output += "\n\tACTIVITY: %s" % self.current.activity
output += "\n\tPOOL: %s" % self.current.pool
output += "\n\tIN EXTERNAL: %s" % self.wf_state['in_external']
output += "\n\tLANE: %s" % self.current.lane_name
output += "\n\tTOKEN: %s" % self.current.token
sys._zops_wf_state_log = output
return output
|
[
"def",
"generate_wf_state_log",
"(",
"self",
")",
":",
"output",
"=",
"'\\n- - - - - -\\n'",
"output",
"+=",
"\"WORKFLOW: %s ( %s )\"",
"%",
"(",
"self",
".",
"current",
".",
"workflow_name",
".",
"upper",
"(",
")",
",",
"self",
".",
"current",
".",
"workflow",
".",
"name",
")",
"output",
"+=",
"\"\\nTASK: %s ( %s )\\n\"",
"%",
"(",
"self",
".",
"current",
".",
"task_name",
",",
"self",
".",
"current",
".",
"task_type",
")",
"output",
"+=",
"\"DATA:\"",
"for",
"k",
",",
"v",
"in",
"self",
".",
"current",
".",
"task_data",
".",
"items",
"(",
")",
":",
"if",
"v",
":",
"output",
"+=",
"\"\\n\\t%s: %s\"",
"%",
"(",
"k",
",",
"v",
")",
"output",
"+=",
"\"\\nCURRENT:\"",
"output",
"+=",
"\"\\n\\tACTIVITY: %s\"",
"%",
"self",
".",
"current",
".",
"activity",
"output",
"+=",
"\"\\n\\tPOOL: %s\"",
"%",
"self",
".",
"current",
".",
"pool",
"output",
"+=",
"\"\\n\\tIN EXTERNAL: %s\"",
"%",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
"output",
"+=",
"\"\\n\\tLANE: %s\"",
"%",
"self",
".",
"current",
".",
"lane_name",
"output",
"+=",
"\"\\n\\tTOKEN: %s\"",
"%",
"self",
".",
"current",
".",
"token",
"sys",
".",
"_zops_wf_state_log",
"=",
"output",
"return",
"output"
] |
Logs the state of workflow and content of task_data.
|
[
"Logs",
"the",
"state",
"of",
"workflow",
"and",
"content",
"of",
"task_data",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L291-L311
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.switch_from_external_to_main_wf
|
def switch_from_external_to_main_wf(self):
"""
Main workflow switcher.
This method recreates main workflow from `main wf` dict which
was set by external workflow swicther previously.
"""
# in external assigned as True in switch_to_external_wf.
# external_wf should finish EndEvent and it's name should be
# also EndEvent for switching again to main wf.
if self.wf_state['in_external'] and self.current.task_type == 'EndEvent' and \
self.current.task_name == 'EndEvent':
# main_wf information was copied in switch_to_external_wf and it takes this information.
main_wf = self.wf_state['main_wf']
# main_wf_name is assigned to current workflow name again.
self.current.workflow_name = main_wf['name']
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# WF knowledge is taken for main wf.
self.workflow_spec = self.get_worfklow_spec()
# WF instance is started again where leave off.
self.workflow = self.deserialize_workflow(main_wf['step'])
# Current WF is this WF instance.
self.current.workflow = self.workflow
# in_external is assigned as False
self.wf_state['in_external'] = False
# finished is assigned as False, because still in progress.
self.wf_state['finished'] = False
# pool info of main_wf is assigned.
self.wf_state['pool'] = main_wf['pool']
self.current.pool = self.wf_state['pool']
# With main_wf is executed.
self.run()
|
python
|
def switch_from_external_to_main_wf(self):
"""
Main workflow switcher.
This method recreates main workflow from `main wf` dict which
was set by external workflow swicther previously.
"""
# in external assigned as True in switch_to_external_wf.
# external_wf should finish EndEvent and it's name should be
# also EndEvent for switching again to main wf.
if self.wf_state['in_external'] and self.current.task_type == 'EndEvent' and \
self.current.task_name == 'EndEvent':
# main_wf information was copied in switch_to_external_wf and it takes this information.
main_wf = self.wf_state['main_wf']
# main_wf_name is assigned to current workflow name again.
self.current.workflow_name = main_wf['name']
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# WF knowledge is taken for main wf.
self.workflow_spec = self.get_worfklow_spec()
# WF instance is started again where leave off.
self.workflow = self.deserialize_workflow(main_wf['step'])
# Current WF is this WF instance.
self.current.workflow = self.workflow
# in_external is assigned as False
self.wf_state['in_external'] = False
# finished is assigned as False, because still in progress.
self.wf_state['finished'] = False
# pool info of main_wf is assigned.
self.wf_state['pool'] = main_wf['pool']
self.current.pool = self.wf_state['pool']
# With main_wf is executed.
self.run()
|
[
"def",
"switch_from_external_to_main_wf",
"(",
"self",
")",
":",
"# in external assigned as True in switch_to_external_wf.",
"# external_wf should finish EndEvent and it's name should be",
"# also EndEvent for switching again to main wf.",
"if",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
"and",
"self",
".",
"current",
".",
"task_type",
"==",
"'EndEvent'",
"and",
"self",
".",
"current",
".",
"task_name",
"==",
"'EndEvent'",
":",
"# main_wf information was copied in switch_to_external_wf and it takes this information.",
"main_wf",
"=",
"self",
".",
"wf_state",
"[",
"'main_wf'",
"]",
"# main_wf_name is assigned to current workflow name again.",
"self",
".",
"current",
".",
"workflow_name",
"=",
"main_wf",
"[",
"'name'",
"]",
"# For external WF, check permission and authentication. But after cleaning current task.",
"self",
".",
"_clear_current_task",
"(",
")",
"# check for auth and perm. current task cleared, do against new workflow_name",
"self",
".",
"check_for_authentication",
"(",
")",
"self",
".",
"check_for_permission",
"(",
")",
"# WF knowledge is taken for main wf.",
"self",
".",
"workflow_spec",
"=",
"self",
".",
"get_worfklow_spec",
"(",
")",
"# WF instance is started again where leave off.",
"self",
".",
"workflow",
"=",
"self",
".",
"deserialize_workflow",
"(",
"main_wf",
"[",
"'step'",
"]",
")",
"# Current WF is this WF instance.",
"self",
".",
"current",
".",
"workflow",
"=",
"self",
".",
"workflow",
"# in_external is assigned as False",
"self",
".",
"wf_state",
"[",
"'in_external'",
"]",
"=",
"False",
"# finished is assigned as False, because still in progress.",
"self",
".",
"wf_state",
"[",
"'finished'",
"]",
"=",
"False",
"# pool info of main_wf is assigned.",
"self",
".",
"wf_state",
"[",
"'pool'",
"]",
"=",
"main_wf",
"[",
"'pool'",
"]",
"self",
".",
"current",
".",
"pool",
"=",
"self",
".",
"wf_state",
"[",
"'pool'",
"]",
"# With main_wf is executed.",
"self",
".",
"run",
"(",
")"
] |
Main workflow switcher.
This method recreates main workflow from `main wf` dict which
was set by external workflow swicther previously.
|
[
"Main",
"workflow",
"switcher",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L316-L365
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.switch_to_external_wf
|
def switch_to_external_wf(self):
"""
External workflow switcher.
This method copies main workflow information into
a temporary dict `main_wf` and makes external workflow
acting as main workflow.
"""
# External WF name should be stated at main wf diagram and type should be service task.
if (self.current.task_type == 'ServiceTask' and
self.current.task.task_spec.type == 'external'):
log.debug("Entering to EXTERNAL WF")
# Main wf information is copied to main_wf.
main_wf = self.wf_state.copy()
# workflow name from main wf diagram is assigned to current workflow name.
# workflow name must be either in task_data with key 'external_wf' or in main diagram's
# topic.
self.current.workflow_name = self.current.task_data.pop('external_wf', False) or self.\
current.task.task_spec.topic
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# wf knowledge is taken for external wf.
self.workflow_spec = self.get_worfklow_spec()
# New WF instance is created for external wf.
self.workflow = self.create_workflow()
# Current WF is this WF instance.
self.current.workflow = self.workflow
# main_wf: main wf information.
# in_external: it states external wf in progress.
# finished: it shows that main wf didn't finish still progress in external wf.
self.wf_state = {'main_wf': main_wf, 'in_external': True, 'finished': False}
|
python
|
def switch_to_external_wf(self):
"""
External workflow switcher.
This method copies main workflow information into
a temporary dict `main_wf` and makes external workflow
acting as main workflow.
"""
# External WF name should be stated at main wf diagram and type should be service task.
if (self.current.task_type == 'ServiceTask' and
self.current.task.task_spec.type == 'external'):
log.debug("Entering to EXTERNAL WF")
# Main wf information is copied to main_wf.
main_wf = self.wf_state.copy()
# workflow name from main wf diagram is assigned to current workflow name.
# workflow name must be either in task_data with key 'external_wf' or in main diagram's
# topic.
self.current.workflow_name = self.current.task_data.pop('external_wf', False) or self.\
current.task.task_spec.topic
# For external WF, check permission and authentication. But after cleaning current task.
self._clear_current_task()
# check for auth and perm. current task cleared, do against new workflow_name
self.check_for_authentication()
self.check_for_permission()
# wf knowledge is taken for external wf.
self.workflow_spec = self.get_worfklow_spec()
# New WF instance is created for external wf.
self.workflow = self.create_workflow()
# Current WF is this WF instance.
self.current.workflow = self.workflow
# main_wf: main wf information.
# in_external: it states external wf in progress.
# finished: it shows that main wf didn't finish still progress in external wf.
self.wf_state = {'main_wf': main_wf, 'in_external': True, 'finished': False}
|
[
"def",
"switch_to_external_wf",
"(",
"self",
")",
":",
"# External WF name should be stated at main wf diagram and type should be service task.",
"if",
"(",
"self",
".",
"current",
".",
"task_type",
"==",
"'ServiceTask'",
"and",
"self",
".",
"current",
".",
"task",
".",
"task_spec",
".",
"type",
"==",
"'external'",
")",
":",
"log",
".",
"debug",
"(",
"\"Entering to EXTERNAL WF\"",
")",
"# Main wf information is copied to main_wf.",
"main_wf",
"=",
"self",
".",
"wf_state",
".",
"copy",
"(",
")",
"# workflow name from main wf diagram is assigned to current workflow name.",
"# workflow name must be either in task_data with key 'external_wf' or in main diagram's",
"# topic.",
"self",
".",
"current",
".",
"workflow_name",
"=",
"self",
".",
"current",
".",
"task_data",
".",
"pop",
"(",
"'external_wf'",
",",
"False",
")",
"or",
"self",
".",
"current",
".",
"task",
".",
"task_spec",
".",
"topic",
"# For external WF, check permission and authentication. But after cleaning current task.",
"self",
".",
"_clear_current_task",
"(",
")",
"# check for auth and perm. current task cleared, do against new workflow_name",
"self",
".",
"check_for_authentication",
"(",
")",
"self",
".",
"check_for_permission",
"(",
")",
"# wf knowledge is taken for external wf.",
"self",
".",
"workflow_spec",
"=",
"self",
".",
"get_worfklow_spec",
"(",
")",
"# New WF instance is created for external wf.",
"self",
".",
"workflow",
"=",
"self",
".",
"create_workflow",
"(",
")",
"# Current WF is this WF instance.",
"self",
".",
"current",
".",
"workflow",
"=",
"self",
".",
"workflow",
"# main_wf: main wf information.",
"# in_external: it states external wf in progress.",
"# finished: it shows that main wf didn't finish still progress in external wf.",
"self",
".",
"wf_state",
"=",
"{",
"'main_wf'",
":",
"main_wf",
",",
"'in_external'",
":",
"True",
",",
"'finished'",
":",
"False",
"}"
] |
External workflow switcher.
This method copies main workflow information into
a temporary dict `main_wf` and makes external workflow
acting as main workflow.
|
[
"External",
"workflow",
"switcher",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L367-L408
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._clear_current_task
|
def _clear_current_task(self):
"""
Clear tasks related attributes, checks permissions
While switching WF to WF, authentication and permissions are checked for new WF.
"""
self.current.task_name = None
self.current.task_type = None
self.current.task = None
|
python
|
def _clear_current_task(self):
"""
Clear tasks related attributes, checks permissions
While switching WF to WF, authentication and permissions are checked for new WF.
"""
self.current.task_name = None
self.current.task_type = None
self.current.task = None
|
[
"def",
"_clear_current_task",
"(",
"self",
")",
":",
"self",
".",
"current",
".",
"task_name",
"=",
"None",
"self",
".",
"current",
".",
"task_type",
"=",
"None",
"self",
".",
"current",
".",
"task",
"=",
"None"
] |
Clear tasks related attributes, checks permissions
While switching WF to WF, authentication and permissions are checked for new WF.
|
[
"Clear",
"tasks",
"related",
"attributes",
"checks",
"permissions",
"While",
"switching",
"WF",
"to",
"WF",
"authentication",
"and",
"permissions",
"are",
"checked",
"for",
"new",
"WF",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L410-L418
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.run
|
def run(self):
"""
Main loop of the workflow engine
- Updates ::class:`~WFCurrent` object.
- Checks for Permissions.
- Activates all READY tasks.
- Runs referenced activities (method calls).
- Saves WF states.
- Stops if current task is a UserTask or EndTask.
- Deletes state object if we finish the WF.
"""
# FIXME: raise if first task after line change isn't a UserTask
# FIXME: raise if last task of a workflow is a UserTask
# actually this check should be done at parser
is_lane_changed = False
while self._should_we_run():
self.check_for_rerun_user_task()
task = None
for task in self.workflow.get_tasks(state=Task.READY):
self.current.old_lane = self.current.lane_name
self.current._update_task(task)
if self.catch_lane_change():
return
self.check_for_permission()
self.check_for_lane_permission()
self.log_wf_state()
self.switch_lang()
self.run_activity()
self.parse_workflow_messages()
self.workflow.complete_task_from_id(self.current.task.id)
self._save_or_delete_workflow()
self.switch_to_external_wf()
if task is None:
break
self.switch_from_external_to_main_wf()
self.current.output['token'] = self.current.token
# look for incoming ready task(s)
for task in self.workflow.get_tasks(state=Task.READY):
self.current._update_task(task)
self.catch_lane_change()
self.handle_wf_finalization()
|
python
|
def run(self):
"""
Main loop of the workflow engine
- Updates ::class:`~WFCurrent` object.
- Checks for Permissions.
- Activates all READY tasks.
- Runs referenced activities (method calls).
- Saves WF states.
- Stops if current task is a UserTask or EndTask.
- Deletes state object if we finish the WF.
"""
# FIXME: raise if first task after line change isn't a UserTask
# FIXME: raise if last task of a workflow is a UserTask
# actually this check should be done at parser
is_lane_changed = False
while self._should_we_run():
self.check_for_rerun_user_task()
task = None
for task in self.workflow.get_tasks(state=Task.READY):
self.current.old_lane = self.current.lane_name
self.current._update_task(task)
if self.catch_lane_change():
return
self.check_for_permission()
self.check_for_lane_permission()
self.log_wf_state()
self.switch_lang()
self.run_activity()
self.parse_workflow_messages()
self.workflow.complete_task_from_id(self.current.task.id)
self._save_or_delete_workflow()
self.switch_to_external_wf()
if task is None:
break
self.switch_from_external_to_main_wf()
self.current.output['token'] = self.current.token
# look for incoming ready task(s)
for task in self.workflow.get_tasks(state=Task.READY):
self.current._update_task(task)
self.catch_lane_change()
self.handle_wf_finalization()
|
[
"def",
"run",
"(",
"self",
")",
":",
"# FIXME: raise if first task after line change isn't a UserTask",
"# FIXME: raise if last task of a workflow is a UserTask",
"# actually this check should be done at parser",
"is_lane_changed",
"=",
"False",
"while",
"self",
".",
"_should_we_run",
"(",
")",
":",
"self",
".",
"check_for_rerun_user_task",
"(",
")",
"task",
"=",
"None",
"for",
"task",
"in",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"state",
"=",
"Task",
".",
"READY",
")",
":",
"self",
".",
"current",
".",
"old_lane",
"=",
"self",
".",
"current",
".",
"lane_name",
"self",
".",
"current",
".",
"_update_task",
"(",
"task",
")",
"if",
"self",
".",
"catch_lane_change",
"(",
")",
":",
"return",
"self",
".",
"check_for_permission",
"(",
")",
"self",
".",
"check_for_lane_permission",
"(",
")",
"self",
".",
"log_wf_state",
"(",
")",
"self",
".",
"switch_lang",
"(",
")",
"self",
".",
"run_activity",
"(",
")",
"self",
".",
"parse_workflow_messages",
"(",
")",
"self",
".",
"workflow",
".",
"complete_task_from_id",
"(",
"self",
".",
"current",
".",
"task",
".",
"id",
")",
"self",
".",
"_save_or_delete_workflow",
"(",
")",
"self",
".",
"switch_to_external_wf",
"(",
")",
"if",
"task",
"is",
"None",
":",
"break",
"self",
".",
"switch_from_external_to_main_wf",
"(",
")",
"self",
".",
"current",
".",
"output",
"[",
"'token'",
"]",
"=",
"self",
".",
"current",
".",
"token",
"# look for incoming ready task(s)",
"for",
"task",
"in",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"state",
"=",
"Task",
".",
"READY",
")",
":",
"self",
".",
"current",
".",
"_update_task",
"(",
"task",
")",
"self",
".",
"catch_lane_change",
"(",
")",
"self",
".",
"handle_wf_finalization",
"(",
")"
] |
Main loop of the workflow engine
- Updates ::class:`~WFCurrent` object.
- Checks for Permissions.
- Activates all READY tasks.
- Runs referenced activities (method calls).
- Saves WF states.
- Stops if current task is a UserTask or EndTask.
- Deletes state object if we finish the WF.
|
[
"Main",
"loop",
"of",
"the",
"workflow",
"engine"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L432-L477
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_rerun_user_task
|
def check_for_rerun_user_task(self):
"""
Checks that the user task needs to re-run.
If necessary, current task and pre task's states are changed and re-run.
If wf_meta not in data(there is no user interaction from pre-task) and last completed task
type is user task and current step is not EndEvent and there is no lane change,
this user task is rerun.
"""
data = self.current.input
if 'wf_meta' in data:
return
current_task = self.workflow.get_tasks(Task.READY)[0]
current_task_type = current_task.task_spec.__class__.__name__
pre_task = current_task.parent
pre_task_type = pre_task.task_spec.__class__.__name__
if pre_task_type != 'UserTask':
return
if current_task_type == 'EndEvent':
return
pre_lane = pre_task.task_spec.lane
current_lane = current_task.task_spec.lane
if pre_lane == current_lane:
pre_task._set_state(Task.READY)
current_task._set_state(Task.MAYBE)
|
python
|
def check_for_rerun_user_task(self):
"""
Checks that the user task needs to re-run.
If necessary, current task and pre task's states are changed and re-run.
If wf_meta not in data(there is no user interaction from pre-task) and last completed task
type is user task and current step is not EndEvent and there is no lane change,
this user task is rerun.
"""
data = self.current.input
if 'wf_meta' in data:
return
current_task = self.workflow.get_tasks(Task.READY)[0]
current_task_type = current_task.task_spec.__class__.__name__
pre_task = current_task.parent
pre_task_type = pre_task.task_spec.__class__.__name__
if pre_task_type != 'UserTask':
return
if current_task_type == 'EndEvent':
return
pre_lane = pre_task.task_spec.lane
current_lane = current_task.task_spec.lane
if pre_lane == current_lane:
pre_task._set_state(Task.READY)
current_task._set_state(Task.MAYBE)
|
[
"def",
"check_for_rerun_user_task",
"(",
"self",
")",
":",
"data",
"=",
"self",
".",
"current",
".",
"input",
"if",
"'wf_meta'",
"in",
"data",
":",
"return",
"current_task",
"=",
"self",
".",
"workflow",
".",
"get_tasks",
"(",
"Task",
".",
"READY",
")",
"[",
"0",
"]",
"current_task_type",
"=",
"current_task",
".",
"task_spec",
".",
"__class__",
".",
"__name__",
"pre_task",
"=",
"current_task",
".",
"parent",
"pre_task_type",
"=",
"pre_task",
".",
"task_spec",
".",
"__class__",
".",
"__name__",
"if",
"pre_task_type",
"!=",
"'UserTask'",
":",
"return",
"if",
"current_task_type",
"==",
"'EndEvent'",
":",
"return",
"pre_lane",
"=",
"pre_task",
".",
"task_spec",
".",
"lane",
"current_lane",
"=",
"current_task",
".",
"task_spec",
".",
"lane",
"if",
"pre_lane",
"==",
"current_lane",
":",
"pre_task",
".",
"_set_state",
"(",
"Task",
".",
"READY",
")",
"current_task",
".",
"_set_state",
"(",
"Task",
".",
"MAYBE",
")"
] |
Checks that the user task needs to re-run.
If necessary, current task and pre task's states are changed and re-run.
If wf_meta not in data(there is no user interaction from pre-task) and last completed task
type is user task and current step is not EndEvent and there is no lane change,
this user task is rerun.
|
[
"Checks",
"that",
"the",
"user",
"task",
"needs",
"to",
"re",
"-",
"run",
".",
"If",
"necessary",
"current",
"task",
"and",
"pre",
"task",
"s",
"states",
"are",
"changed",
"and",
"re",
"-",
"run",
".",
"If",
"wf_meta",
"not",
"in",
"data",
"(",
"there",
"is",
"no",
"user",
"interaction",
"from",
"pre",
"-",
"task",
")",
"and",
"last",
"completed",
"task",
"type",
"is",
"user",
"task",
"and",
"current",
"step",
"is",
"not",
"EndEvent",
"and",
"there",
"is",
"no",
"lane",
"change",
"this",
"user",
"task",
"is",
"rerun",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L479-L506
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.switch_lang
|
def switch_lang(self):
"""Switch to the language of the current user.
If the current language is already the specified one, nothing will be done.
"""
locale = self.current.locale
translation.InstalledLocale.install_language(locale['locale_language'])
translation.InstalledLocale.install_locale(locale['locale_datetime'], 'datetime')
translation.InstalledLocale.install_locale(locale['locale_number'], 'number')
|
python
|
def switch_lang(self):
"""Switch to the language of the current user.
If the current language is already the specified one, nothing will be done.
"""
locale = self.current.locale
translation.InstalledLocale.install_language(locale['locale_language'])
translation.InstalledLocale.install_locale(locale['locale_datetime'], 'datetime')
translation.InstalledLocale.install_locale(locale['locale_number'], 'number')
|
[
"def",
"switch_lang",
"(",
"self",
")",
":",
"locale",
"=",
"self",
".",
"current",
".",
"locale",
"translation",
".",
"InstalledLocale",
".",
"install_language",
"(",
"locale",
"[",
"'locale_language'",
"]",
")",
"translation",
".",
"InstalledLocale",
".",
"install_locale",
"(",
"locale",
"[",
"'locale_datetime'",
"]",
",",
"'datetime'",
")",
"translation",
".",
"InstalledLocale",
".",
"install_locale",
"(",
"locale",
"[",
"'locale_number'",
"]",
",",
"'number'",
")"
] |
Switch to the language of the current user.
If the current language is already the specified one, nothing will be done.
|
[
"Switch",
"to",
"the",
"language",
"of",
"the",
"current",
"user",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L508-L516
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.catch_lane_change
|
def catch_lane_change(self):
"""
trigger a lane_user_change signal if we switched to a new lane
and new lane's user is different from current one
"""
if self.current.lane_name:
if self.current.old_lane and self.current.lane_name != self.current.old_lane:
# if lane_name not found in pool or it's user different from the current(old) user
if (self.current.lane_id not in self.current.pool or
self.current.pool[self.current.lane_id] != self.current.user_id):
self.current.log.info("LANE CHANGE : %s >> %s" % (self.current.old_lane,
self.current.lane_name))
if self.current.lane_auto_sendoff:
self.current.sendoff_current_user()
self.current.flow_enabled = False
if self.current.lane_auto_invite:
self.current.invite_other_parties(self._get_possible_lane_owners())
return True
|
python
|
def catch_lane_change(self):
"""
trigger a lane_user_change signal if we switched to a new lane
and new lane's user is different from current one
"""
if self.current.lane_name:
if self.current.old_lane and self.current.lane_name != self.current.old_lane:
# if lane_name not found in pool or it's user different from the current(old) user
if (self.current.lane_id not in self.current.pool or
self.current.pool[self.current.lane_id] != self.current.user_id):
self.current.log.info("LANE CHANGE : %s >> %s" % (self.current.old_lane,
self.current.lane_name))
if self.current.lane_auto_sendoff:
self.current.sendoff_current_user()
self.current.flow_enabled = False
if self.current.lane_auto_invite:
self.current.invite_other_parties(self._get_possible_lane_owners())
return True
|
[
"def",
"catch_lane_change",
"(",
"self",
")",
":",
"if",
"self",
".",
"current",
".",
"lane_name",
":",
"if",
"self",
".",
"current",
".",
"old_lane",
"and",
"self",
".",
"current",
".",
"lane_name",
"!=",
"self",
".",
"current",
".",
"old_lane",
":",
"# if lane_name not found in pool or it's user different from the current(old) user",
"if",
"(",
"self",
".",
"current",
".",
"lane_id",
"not",
"in",
"self",
".",
"current",
".",
"pool",
"or",
"self",
".",
"current",
".",
"pool",
"[",
"self",
".",
"current",
".",
"lane_id",
"]",
"!=",
"self",
".",
"current",
".",
"user_id",
")",
":",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"LANE CHANGE : %s >> %s\"",
"%",
"(",
"self",
".",
"current",
".",
"old_lane",
",",
"self",
".",
"current",
".",
"lane_name",
")",
")",
"if",
"self",
".",
"current",
".",
"lane_auto_sendoff",
":",
"self",
".",
"current",
".",
"sendoff_current_user",
"(",
")",
"self",
".",
"current",
".",
"flow_enabled",
"=",
"False",
"if",
"self",
".",
"current",
".",
"lane_auto_invite",
":",
"self",
".",
"current",
".",
"invite_other_parties",
"(",
"self",
".",
"_get_possible_lane_owners",
"(",
")",
")",
"return",
"True"
] |
trigger a lane_user_change signal if we switched to a new lane
and new lane's user is different from current one
|
[
"trigger",
"a",
"lane_user_change",
"signal",
"if",
"we",
"switched",
"to",
"a",
"new",
"lane",
"and",
"new",
"lane",
"s",
"user",
"is",
"different",
"from",
"current",
"one"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L518-L535
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.parse_workflow_messages
|
def parse_workflow_messages(self):
"""
Transmits client message that defined in
a workflow task's inputOutput extension
.. code-block:: xml
<bpmn2:extensionElements>
<camunda:inputOutput>
<camunda:inputParameter name="client_message">
<camunda:map>
<camunda:entry key="title">Teşekkürler</camunda:entry>
<camunda:entry key="body">İşlem Başarılı</camunda:entry>
<camunda:entry key="type">info</camunda:entry>
</camunda:map>
</camunda:inputParameter>
</camunda:inputOutput>
</bpmn2:extensionElements>
"""
if 'client_message' in self.current.spec.data:
m = self.current.spec.data['client_message']
self.current.msg_box(title=m.get('title'),
msg=m.get('body'),
typ=m.get('type', 'info'))
|
python
|
def parse_workflow_messages(self):
"""
Transmits client message that defined in
a workflow task's inputOutput extension
.. code-block:: xml
<bpmn2:extensionElements>
<camunda:inputOutput>
<camunda:inputParameter name="client_message">
<camunda:map>
<camunda:entry key="title">Teşekkürler</camunda:entry>
<camunda:entry key="body">İşlem Başarılı</camunda:entry>
<camunda:entry key="type">info</camunda:entry>
</camunda:map>
</camunda:inputParameter>
</camunda:inputOutput>
</bpmn2:extensionElements>
"""
if 'client_message' in self.current.spec.data:
m = self.current.spec.data['client_message']
self.current.msg_box(title=m.get('title'),
msg=m.get('body'),
typ=m.get('type', 'info'))
|
[
"def",
"parse_workflow_messages",
"(",
"self",
")",
":",
"if",
"'client_message'",
"in",
"self",
".",
"current",
".",
"spec",
".",
"data",
":",
"m",
"=",
"self",
".",
"current",
".",
"spec",
".",
"data",
"[",
"'client_message'",
"]",
"self",
".",
"current",
".",
"msg_box",
"(",
"title",
"=",
"m",
".",
"get",
"(",
"'title'",
")",
",",
"msg",
"=",
"m",
".",
"get",
"(",
"'body'",
")",
",",
"typ",
"=",
"m",
".",
"get",
"(",
"'type'",
",",
"'info'",
")",
")"
] |
Transmits client message that defined in
a workflow task's inputOutput extension
.. code-block:: xml
<bpmn2:extensionElements>
<camunda:inputOutput>
<camunda:inputParameter name="client_message">
<camunda:map>
<camunda:entry key="title">Teşekkürler</camunda:entry>
<camunda:entry key="body">İşlem Başarılı</camunda:entry>
<camunda:entry key="type">info</camunda:entry>
</camunda:map>
</camunda:inputParameter>
</camunda:inputOutput>
</bpmn2:extensionElements>
|
[
"Transmits",
"client",
"message",
"that",
"defined",
"in",
"a",
"workflow",
"task",
"s",
"inputOutput",
"extension"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L538-L562
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.run_activity
|
def run_activity(self):
"""
runs the method that referenced from current task
"""
activity = self.current.activity
if activity:
if activity not in self.wf_activities:
self._load_activity(activity)
self.current.log.debug(
"Calling Activity %s from %s" % (activity, self.wf_activities[activity]))
self.wf_activities[self.current.activity](self.current)
|
python
|
def run_activity(self):
"""
runs the method that referenced from current task
"""
activity = self.current.activity
if activity:
if activity not in self.wf_activities:
self._load_activity(activity)
self.current.log.debug(
"Calling Activity %s from %s" % (activity, self.wf_activities[activity]))
self.wf_activities[self.current.activity](self.current)
|
[
"def",
"run_activity",
"(",
"self",
")",
":",
"activity",
"=",
"self",
".",
"current",
".",
"activity",
"if",
"activity",
":",
"if",
"activity",
"not",
"in",
"self",
".",
"wf_activities",
":",
"self",
".",
"_load_activity",
"(",
"activity",
")",
"self",
".",
"current",
".",
"log",
".",
"debug",
"(",
"\"Calling Activity %s from %s\"",
"%",
"(",
"activity",
",",
"self",
".",
"wf_activities",
"[",
"activity",
"]",
")",
")",
"self",
".",
"wf_activities",
"[",
"self",
".",
"current",
".",
"activity",
"]",
"(",
"self",
".",
"current",
")"
] |
runs the method that referenced from current task
|
[
"runs",
"the",
"method",
"that",
"referenced",
"from",
"current",
"task"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L573-L583
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._import_object
|
def _import_object(self, path, look_for_cls_method):
"""
Imports the module that contains the referenced method.
Args:
path: python path of class/function
look_for_cls_method (bool): If True, treat the last part of path as class method.
Returns:
Tuple. (class object, class name, method to be called)
"""
last_nth = 2 if look_for_cls_method else 1
path = path.split('.')
module_path = '.'.join(path[:-last_nth])
class_name = path[-last_nth]
module = importlib.import_module(module_path)
if look_for_cls_method and path[-last_nth:][0] == path[-last_nth]:
class_method = path[-last_nth:][1]
else:
class_method = None
return getattr(module, class_name), class_name, class_method
|
python
|
def _import_object(self, path, look_for_cls_method):
"""
Imports the module that contains the referenced method.
Args:
path: python path of class/function
look_for_cls_method (bool): If True, treat the last part of path as class method.
Returns:
Tuple. (class object, class name, method to be called)
"""
last_nth = 2 if look_for_cls_method else 1
path = path.split('.')
module_path = '.'.join(path[:-last_nth])
class_name = path[-last_nth]
module = importlib.import_module(module_path)
if look_for_cls_method and path[-last_nth:][0] == path[-last_nth]:
class_method = path[-last_nth:][1]
else:
class_method = None
return getattr(module, class_name), class_name, class_method
|
[
"def",
"_import_object",
"(",
"self",
",",
"path",
",",
"look_for_cls_method",
")",
":",
"last_nth",
"=",
"2",
"if",
"look_for_cls_method",
"else",
"1",
"path",
"=",
"path",
".",
"split",
"(",
"'.'",
")",
"module_path",
"=",
"'.'",
".",
"join",
"(",
"path",
"[",
":",
"-",
"last_nth",
"]",
")",
"class_name",
"=",
"path",
"[",
"-",
"last_nth",
"]",
"module",
"=",
"importlib",
".",
"import_module",
"(",
"module_path",
")",
"if",
"look_for_cls_method",
"and",
"path",
"[",
"-",
"last_nth",
":",
"]",
"[",
"0",
"]",
"==",
"path",
"[",
"-",
"last_nth",
"]",
":",
"class_method",
"=",
"path",
"[",
"-",
"last_nth",
":",
"]",
"[",
"1",
"]",
"else",
":",
"class_method",
"=",
"None",
"return",
"getattr",
"(",
"module",
",",
"class_name",
")",
",",
"class_name",
",",
"class_method"
] |
Imports the module that contains the referenced method.
Args:
path: python path of class/function
look_for_cls_method (bool): If True, treat the last part of path as class method.
Returns:
Tuple. (class object, class name, method to be called)
|
[
"Imports",
"the",
"module",
"that",
"contains",
"the",
"referenced",
"method",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L585-L606
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine._load_activity
|
def _load_activity(self, activity):
"""
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
"""
fpths = []
full_path = ''
errors = []
paths = settings.ACTIVITY_MODULES_IMPORT_PATHS
number_of_paths = len(paths)
for index_no in range(number_of_paths):
full_path = "%s.%s" % (paths[index_no], activity)
for look4kls in (0, 1):
try:
self.current.log.info("try to load from %s[%s]" % (full_path, look4kls))
kls, cls_name, cls_method = self._import_object(full_path, look4kls)
if cls_method:
self.current.log.info("WILLCall %s(current).%s()" % (kls, cls_method))
self.wf_activities[activity] = lambda crnt: getattr(kls(crnt), cls_method)()
else:
self.wf_activities[activity] = kls
return
except (ImportError, AttributeError):
fpths.append(full_path)
errmsg = "{activity} not found under these paths:\n\n >>> {paths} \n\n" \
"Error Messages:\n {errors}"
errors.append("\n========================================================>\n"
"| PATH | %s"
"\n========================================================>\n\n"
"%s" % (full_path, traceback.format_exc()))
assert index_no != number_of_paths - 1, errmsg.format(activity=activity,
paths='\n >>> '.join(
set(fpths)),
errors='\n\n'.join(errors)
)
except:
self.current.log.exception("Cannot found the %s" % activity)
|
python
|
def _load_activity(self, activity):
"""
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
"""
fpths = []
full_path = ''
errors = []
paths = settings.ACTIVITY_MODULES_IMPORT_PATHS
number_of_paths = len(paths)
for index_no in range(number_of_paths):
full_path = "%s.%s" % (paths[index_no], activity)
for look4kls in (0, 1):
try:
self.current.log.info("try to load from %s[%s]" % (full_path, look4kls))
kls, cls_name, cls_method = self._import_object(full_path, look4kls)
if cls_method:
self.current.log.info("WILLCall %s(current).%s()" % (kls, cls_method))
self.wf_activities[activity] = lambda crnt: getattr(kls(crnt), cls_method)()
else:
self.wf_activities[activity] = kls
return
except (ImportError, AttributeError):
fpths.append(full_path)
errmsg = "{activity} not found under these paths:\n\n >>> {paths} \n\n" \
"Error Messages:\n {errors}"
errors.append("\n========================================================>\n"
"| PATH | %s"
"\n========================================================>\n\n"
"%s" % (full_path, traceback.format_exc()))
assert index_no != number_of_paths - 1, errmsg.format(activity=activity,
paths='\n >>> '.join(
set(fpths)),
errors='\n\n'.join(errors)
)
except:
self.current.log.exception("Cannot found the %s" % activity)
|
[
"def",
"_load_activity",
"(",
"self",
",",
"activity",
")",
":",
"fpths",
"=",
"[",
"]",
"full_path",
"=",
"''",
"errors",
"=",
"[",
"]",
"paths",
"=",
"settings",
".",
"ACTIVITY_MODULES_IMPORT_PATHS",
"number_of_paths",
"=",
"len",
"(",
"paths",
")",
"for",
"index_no",
"in",
"range",
"(",
"number_of_paths",
")",
":",
"full_path",
"=",
"\"%s.%s\"",
"%",
"(",
"paths",
"[",
"index_no",
"]",
",",
"activity",
")",
"for",
"look4kls",
"in",
"(",
"0",
",",
"1",
")",
":",
"try",
":",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"try to load from %s[%s]\"",
"%",
"(",
"full_path",
",",
"look4kls",
")",
")",
"kls",
",",
"cls_name",
",",
"cls_method",
"=",
"self",
".",
"_import_object",
"(",
"full_path",
",",
"look4kls",
")",
"if",
"cls_method",
":",
"self",
".",
"current",
".",
"log",
".",
"info",
"(",
"\"WILLCall %s(current).%s()\"",
"%",
"(",
"kls",
",",
"cls_method",
")",
")",
"self",
".",
"wf_activities",
"[",
"activity",
"]",
"=",
"lambda",
"crnt",
":",
"getattr",
"(",
"kls",
"(",
"crnt",
")",
",",
"cls_method",
")",
"(",
")",
"else",
":",
"self",
".",
"wf_activities",
"[",
"activity",
"]",
"=",
"kls",
"return",
"except",
"(",
"ImportError",
",",
"AttributeError",
")",
":",
"fpths",
".",
"append",
"(",
"full_path",
")",
"errmsg",
"=",
"\"{activity} not found under these paths:\\n\\n >>> {paths} \\n\\n\"",
"\"Error Messages:\\n {errors}\"",
"errors",
".",
"append",
"(",
"\"\\n========================================================>\\n\"",
"\"| PATH | %s\"",
"\"\\n========================================================>\\n\\n\"",
"\"%s\"",
"%",
"(",
"full_path",
",",
"traceback",
".",
"format_exc",
"(",
")",
")",
")",
"assert",
"index_no",
"!=",
"number_of_paths",
"-",
"1",
",",
"errmsg",
".",
"format",
"(",
"activity",
"=",
"activity",
",",
"paths",
"=",
"'\\n >>> '",
".",
"join",
"(",
"set",
"(",
"fpths",
")",
")",
",",
"errors",
"=",
"'\\n\\n'",
".",
"join",
"(",
"errors",
")",
")",
"except",
":",
"self",
".",
"current",
".",
"log",
".",
"exception",
"(",
"\"Cannot found the %s\"",
"%",
"activity",
")"
] |
Iterates trough the all enabled `~zengine.settings.ACTIVITY_MODULES_IMPORT_PATHS` to find the given path.
|
[
"Iterates",
"trough",
"the",
"all",
"enabled",
"~zengine",
".",
"settings",
".",
"ACTIVITY_MODULES_IMPORT_PATHS",
"to",
"find",
"the",
"given",
"path",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L608-L643
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_authentication
|
def check_for_authentication(self):
"""
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list.
Raises:
HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
"""
auth_required = self.current.workflow_name not in settings.ANONYMOUS_WORKFLOWS
if auth_required and not self.current.is_auth:
self.current.log.debug("LOGIN REQUIRED:::: %s" % self.current.workflow_name)
raise HTTPError(401, "Login required for %s" % self.current.workflow_name)
|
python
|
def check_for_authentication(self):
"""
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list.
Raises:
HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
"""
auth_required = self.current.workflow_name not in settings.ANONYMOUS_WORKFLOWS
if auth_required and not self.current.is_auth:
self.current.log.debug("LOGIN REQUIRED:::: %s" % self.current.workflow_name)
raise HTTPError(401, "Login required for %s" % self.current.workflow_name)
|
[
"def",
"check_for_authentication",
"(",
"self",
")",
":",
"auth_required",
"=",
"self",
".",
"current",
".",
"workflow_name",
"not",
"in",
"settings",
".",
"ANONYMOUS_WORKFLOWS",
"if",
"auth_required",
"and",
"not",
"self",
".",
"current",
".",
"is_auth",
":",
"self",
".",
"current",
".",
"log",
".",
"debug",
"(",
"\"LOGIN REQUIRED:::: %s\"",
"%",
"self",
".",
"current",
".",
"workflow_name",
")",
"raise",
"HTTPError",
"(",
"401",
",",
"\"Login required for %s\"",
"%",
"self",
".",
"current",
".",
"workflow_name",
")"
] |
Checks current workflow against :py:data:`~zengine.settings.ANONYMOUS_WORKFLOWS` list.
Raises:
HTTPUnauthorized: if WF needs an authenticated user and current user isn't.
|
[
"Checks",
"current",
"workflow",
"against",
":",
"py",
":",
"data",
":",
"~zengine",
".",
"settings",
".",
"ANONYMOUS_WORKFLOWS",
"list",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L645-L655
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_lane_permission
|
def check_for_lane_permission(self):
"""
One or more permissions can be associated with a lane
of a workflow. In a similar way, a lane can be
restricted with relation to other lanes of the workflow.
This method called on lane changes and checks user has
required permissions and relations.
Raises:
HTTPForbidden: if the current user hasn't got the
required permissions and proper relations
"""
# TODO: Cache lane_data in app memory
if self.current.lane_permission:
log.debug("HAS LANE PERM: %s" % self.current.lane_permission)
perm = self.current.lane_permission
if not self.current.has_permission(perm):
raise HTTPError(403, "You don't have required lane permission: %s" % perm)
if self.current.lane_relations:
context = self.get_pool_context()
log.debug("HAS LANE RELS: %s" % self.current.lane_relations)
try:
cond_result = eval(self.current.lane_relations, context)
except:
log.exception("CONDITION EVAL ERROR : %s || %s" % (
self.current.lane_relations, context))
raise
if not cond_result:
log.debug("LANE RELATION ERR: %s %s" % (self.current.lane_relations, context))
raise HTTPError(403, "You aren't qualified for this lane: %s" %
self.current.lane_relations)
|
python
|
def check_for_lane_permission(self):
"""
One or more permissions can be associated with a lane
of a workflow. In a similar way, a lane can be
restricted with relation to other lanes of the workflow.
This method called on lane changes and checks user has
required permissions and relations.
Raises:
HTTPForbidden: if the current user hasn't got the
required permissions and proper relations
"""
# TODO: Cache lane_data in app memory
if self.current.lane_permission:
log.debug("HAS LANE PERM: %s" % self.current.lane_permission)
perm = self.current.lane_permission
if not self.current.has_permission(perm):
raise HTTPError(403, "You don't have required lane permission: %s" % perm)
if self.current.lane_relations:
context = self.get_pool_context()
log.debug("HAS LANE RELS: %s" % self.current.lane_relations)
try:
cond_result = eval(self.current.lane_relations, context)
except:
log.exception("CONDITION EVAL ERROR : %s || %s" % (
self.current.lane_relations, context))
raise
if not cond_result:
log.debug("LANE RELATION ERR: %s %s" % (self.current.lane_relations, context))
raise HTTPError(403, "You aren't qualified for this lane: %s" %
self.current.lane_relations)
|
[
"def",
"check_for_lane_permission",
"(",
"self",
")",
":",
"# TODO: Cache lane_data in app memory",
"if",
"self",
".",
"current",
".",
"lane_permission",
":",
"log",
".",
"debug",
"(",
"\"HAS LANE PERM: %s\"",
"%",
"self",
".",
"current",
".",
"lane_permission",
")",
"perm",
"=",
"self",
".",
"current",
".",
"lane_permission",
"if",
"not",
"self",
".",
"current",
".",
"has_permission",
"(",
"perm",
")",
":",
"raise",
"HTTPError",
"(",
"403",
",",
"\"You don't have required lane permission: %s\"",
"%",
"perm",
")",
"if",
"self",
".",
"current",
".",
"lane_relations",
":",
"context",
"=",
"self",
".",
"get_pool_context",
"(",
")",
"log",
".",
"debug",
"(",
"\"HAS LANE RELS: %s\"",
"%",
"self",
".",
"current",
".",
"lane_relations",
")",
"try",
":",
"cond_result",
"=",
"eval",
"(",
"self",
".",
"current",
".",
"lane_relations",
",",
"context",
")",
"except",
":",
"log",
".",
"exception",
"(",
"\"CONDITION EVAL ERROR : %s || %s\"",
"%",
"(",
"self",
".",
"current",
".",
"lane_relations",
",",
"context",
")",
")",
"raise",
"if",
"not",
"cond_result",
":",
"log",
".",
"debug",
"(",
"\"LANE RELATION ERR: %s %s\"",
"%",
"(",
"self",
".",
"current",
".",
"lane_relations",
",",
"context",
")",
")",
"raise",
"HTTPError",
"(",
"403",
",",
"\"You aren't qualified for this lane: %s\"",
"%",
"self",
".",
"current",
".",
"lane_relations",
")"
] |
One or more permissions can be associated with a lane
of a workflow. In a similar way, a lane can be
restricted with relation to other lanes of the workflow.
This method called on lane changes and checks user has
required permissions and relations.
Raises:
HTTPForbidden: if the current user hasn't got the
required permissions and proper relations
|
[
"One",
"or",
"more",
"permissions",
"can",
"be",
"associated",
"with",
"a",
"lane",
"of",
"a",
"workflow",
".",
"In",
"a",
"similar",
"way",
"a",
"lane",
"can",
"be",
"restricted",
"with",
"relation",
"to",
"other",
"lanes",
"of",
"the",
"workflow",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L657-L690
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.check_for_permission
|
def check_for_permission(self):
# TODO: Works but not beautiful, needs review!
"""
Checks if current user (or role) has the required permission
for current workflow step.
Raises:
HTTPError: if user doesn't have required permissions.
"""
if self.current.task:
lane = self.current.lane_id
permission = "%s.%s.%s" % (self.current.workflow_name, lane, self.current.task_name)
else:
permission = self.current.workflow_name
log.debug("CHECK PERM: %s" % permission)
if (self.current.task_type not in PERM_REQ_TASK_TYPES or
permission.startswith(tuple(settings.ANONYMOUS_WORKFLOWS)) or
(self.current.is_auth and permission.startswith(tuple(settings.COMMON_WORKFLOWS)))):
return
# FIXME:needs hardening
log.debug("REQUIRE PERM: %s" % permission)
if not self.current.has_permission(permission):
raise HTTPError(403, "You don't have required permission: %s" % permission)
|
python
|
def check_for_permission(self):
# TODO: Works but not beautiful, needs review!
"""
Checks if current user (or role) has the required permission
for current workflow step.
Raises:
HTTPError: if user doesn't have required permissions.
"""
if self.current.task:
lane = self.current.lane_id
permission = "%s.%s.%s" % (self.current.workflow_name, lane, self.current.task_name)
else:
permission = self.current.workflow_name
log.debug("CHECK PERM: %s" % permission)
if (self.current.task_type not in PERM_REQ_TASK_TYPES or
permission.startswith(tuple(settings.ANONYMOUS_WORKFLOWS)) or
(self.current.is_auth and permission.startswith(tuple(settings.COMMON_WORKFLOWS)))):
return
# FIXME:needs hardening
log.debug("REQUIRE PERM: %s" % permission)
if not self.current.has_permission(permission):
raise HTTPError(403, "You don't have required permission: %s" % permission)
|
[
"def",
"check_for_permission",
"(",
"self",
")",
":",
"# TODO: Works but not beautiful, needs review!",
"if",
"self",
".",
"current",
".",
"task",
":",
"lane",
"=",
"self",
".",
"current",
".",
"lane_id",
"permission",
"=",
"\"%s.%s.%s\"",
"%",
"(",
"self",
".",
"current",
".",
"workflow_name",
",",
"lane",
",",
"self",
".",
"current",
".",
"task_name",
")",
"else",
":",
"permission",
"=",
"self",
".",
"current",
".",
"workflow_name",
"log",
".",
"debug",
"(",
"\"CHECK PERM: %s\"",
"%",
"permission",
")",
"if",
"(",
"self",
".",
"current",
".",
"task_type",
"not",
"in",
"PERM_REQ_TASK_TYPES",
"or",
"permission",
".",
"startswith",
"(",
"tuple",
"(",
"settings",
".",
"ANONYMOUS_WORKFLOWS",
")",
")",
"or",
"(",
"self",
".",
"current",
".",
"is_auth",
"and",
"permission",
".",
"startswith",
"(",
"tuple",
"(",
"settings",
".",
"COMMON_WORKFLOWS",
")",
")",
")",
")",
":",
"return",
"# FIXME:needs hardening",
"log",
".",
"debug",
"(",
"\"REQUIRE PERM: %s\"",
"%",
"permission",
")",
"if",
"not",
"self",
".",
"current",
".",
"has_permission",
"(",
"permission",
")",
":",
"raise",
"HTTPError",
"(",
"403",
",",
"\"You don't have required permission: %s\"",
"%",
"permission",
")"
] |
Checks if current user (or role) has the required permission
for current workflow step.
Raises:
HTTPError: if user doesn't have required permissions.
|
[
"Checks",
"if",
"current",
"user",
"(",
"or",
"role",
")",
"has",
"the",
"required",
"permission",
"for",
"current",
"workflow",
"step",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L692-L716
|
zetaops/zengine
|
zengine/engine.py
|
ZEngine.handle_wf_finalization
|
def handle_wf_finalization(self):
"""
Removes the ``token`` key from ``current.output`` if WF is over.
"""
if ((not self.current.flow_enabled or (
self.current.task_type.startswith('End') and not self.are_we_in_subprocess())) and
'token' in self.current.output):
del self.current.output['token']
|
python
|
def handle_wf_finalization(self):
"""
Removes the ``token`` key from ``current.output`` if WF is over.
"""
if ((not self.current.flow_enabled or (
self.current.task_type.startswith('End') and not self.are_we_in_subprocess())) and
'token' in self.current.output):
del self.current.output['token']
|
[
"def",
"handle_wf_finalization",
"(",
"self",
")",
":",
"if",
"(",
"(",
"not",
"self",
".",
"current",
".",
"flow_enabled",
"or",
"(",
"self",
".",
"current",
".",
"task_type",
".",
"startswith",
"(",
"'End'",
")",
"and",
"not",
"self",
".",
"are_we_in_subprocess",
"(",
")",
")",
")",
"and",
"'token'",
"in",
"self",
".",
"current",
".",
"output",
")",
":",
"del",
"self",
".",
"current",
".",
"output",
"[",
"'token'",
"]"
] |
Removes the ``token`` key from ``current.output`` if WF is over.
|
[
"Removes",
"the",
"token",
"key",
"from",
"current",
".",
"output",
"if",
"WF",
"is",
"over",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/engine.py#L718-L725
|
cimm-kzn/CGRtools
|
CGRtools/utils/rdkit.py
|
from_rdkit_molecule
|
def from_rdkit_molecule(data):
"""
RDKit molecule object to MoleculeContainer converter
"""
m = MoleculeContainer()
atoms, mapping = [], []
for a in data.GetAtoms():
atom = {'element': a.GetSymbol(), 'charge': a.GetFormalCharge()}
atoms.append(atom)
mapping.append(a.GetAtomMapNum())
isotope = a.GetIsotope()
if isotope:
atom['isotope'] = isotope
radical = a.GetNumRadicalElectrons()
if radical:
atom['multiplicity'] = radical + 1
conformers = data.GetConformers()
if conformers:
for atom, (x, y, z) in zip(atoms, conformers[0].GetPositions()):
atom['x'] = x
atom['y'] = y
atom['z'] = z
for atom, mapping in zip(atoms, mapping):
a = m.add_atom(atom)
if mapping:
m.atom(a)._parsed_mapping = mapping
for bond in data.GetBonds():
m.add_bond(bond.GetBeginAtomIdx() + 1, bond.GetEndAtomIdx() + 1, _rdkit_bond_map[bond.GetBondType()])
return m
|
python
|
def from_rdkit_molecule(data):
"""
RDKit molecule object to MoleculeContainer converter
"""
m = MoleculeContainer()
atoms, mapping = [], []
for a in data.GetAtoms():
atom = {'element': a.GetSymbol(), 'charge': a.GetFormalCharge()}
atoms.append(atom)
mapping.append(a.GetAtomMapNum())
isotope = a.GetIsotope()
if isotope:
atom['isotope'] = isotope
radical = a.GetNumRadicalElectrons()
if radical:
atom['multiplicity'] = radical + 1
conformers = data.GetConformers()
if conformers:
for atom, (x, y, z) in zip(atoms, conformers[0].GetPositions()):
atom['x'] = x
atom['y'] = y
atom['z'] = z
for atom, mapping in zip(atoms, mapping):
a = m.add_atom(atom)
if mapping:
m.atom(a)._parsed_mapping = mapping
for bond in data.GetBonds():
m.add_bond(bond.GetBeginAtomIdx() + 1, bond.GetEndAtomIdx() + 1, _rdkit_bond_map[bond.GetBondType()])
return m
|
[
"def",
"from_rdkit_molecule",
"(",
"data",
")",
":",
"m",
"=",
"MoleculeContainer",
"(",
")",
"atoms",
",",
"mapping",
"=",
"[",
"]",
",",
"[",
"]",
"for",
"a",
"in",
"data",
".",
"GetAtoms",
"(",
")",
":",
"atom",
"=",
"{",
"'element'",
":",
"a",
".",
"GetSymbol",
"(",
")",
",",
"'charge'",
":",
"a",
".",
"GetFormalCharge",
"(",
")",
"}",
"atoms",
".",
"append",
"(",
"atom",
")",
"mapping",
".",
"append",
"(",
"a",
".",
"GetAtomMapNum",
"(",
")",
")",
"isotope",
"=",
"a",
".",
"GetIsotope",
"(",
")",
"if",
"isotope",
":",
"atom",
"[",
"'isotope'",
"]",
"=",
"isotope",
"radical",
"=",
"a",
".",
"GetNumRadicalElectrons",
"(",
")",
"if",
"radical",
":",
"atom",
"[",
"'multiplicity'",
"]",
"=",
"radical",
"+",
"1",
"conformers",
"=",
"data",
".",
"GetConformers",
"(",
")",
"if",
"conformers",
":",
"for",
"atom",
",",
"(",
"x",
",",
"y",
",",
"z",
")",
"in",
"zip",
"(",
"atoms",
",",
"conformers",
"[",
"0",
"]",
".",
"GetPositions",
"(",
")",
")",
":",
"atom",
"[",
"'x'",
"]",
"=",
"x",
"atom",
"[",
"'y'",
"]",
"=",
"y",
"atom",
"[",
"'z'",
"]",
"=",
"z",
"for",
"atom",
",",
"mapping",
"in",
"zip",
"(",
"atoms",
",",
"mapping",
")",
":",
"a",
"=",
"m",
".",
"add_atom",
"(",
"atom",
")",
"if",
"mapping",
":",
"m",
".",
"atom",
"(",
"a",
")",
".",
"_parsed_mapping",
"=",
"mapping",
"for",
"bond",
"in",
"data",
".",
"GetBonds",
"(",
")",
":",
"m",
".",
"add_bond",
"(",
"bond",
".",
"GetBeginAtomIdx",
"(",
")",
"+",
"1",
",",
"bond",
".",
"GetEndAtomIdx",
"(",
")",
"+",
"1",
",",
"_rdkit_bond_map",
"[",
"bond",
".",
"GetBondType",
"(",
")",
"]",
")",
"return",
"m"
] |
RDKit molecule object to MoleculeContainer converter
|
[
"RDKit",
"molecule",
"object",
"to",
"MoleculeContainer",
"converter"
] |
train
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/utils/rdkit.py#L23-L56
|
cimm-kzn/CGRtools
|
CGRtools/utils/rdkit.py
|
to_rdkit_molecule
|
def to_rdkit_molecule(data):
"""
MoleculeContainer to RDKit molecule object converter
"""
mol = RWMol()
conf = Conformer()
mapping = {}
is_3d = False
for n, a in data.atoms():
ra = Atom(a.number)
ra.SetAtomMapNum(n)
if a.charge:
ra.SetFormalCharge(a.charge)
if a.isotope != a.common_isotope:
ra.SetIsotope(a.isotope)
if a.radical:
ra.SetNumRadicalElectrons(a.radical)
mapping[n] = m = mol.AddAtom(ra)
conf.SetAtomPosition(m, (a.x, a.y, a.z))
if a.z:
is_3d = True
if not is_3d:
conf.Set3D(False)
for n, m, b in data.bonds():
mol.AddBond(mapping[n], mapping[m], _bond_map[b.order])
mol.AddConformer(conf)
SanitizeMol(mol)
return mol
|
python
|
def to_rdkit_molecule(data):
"""
MoleculeContainer to RDKit molecule object converter
"""
mol = RWMol()
conf = Conformer()
mapping = {}
is_3d = False
for n, a in data.atoms():
ra = Atom(a.number)
ra.SetAtomMapNum(n)
if a.charge:
ra.SetFormalCharge(a.charge)
if a.isotope != a.common_isotope:
ra.SetIsotope(a.isotope)
if a.radical:
ra.SetNumRadicalElectrons(a.radical)
mapping[n] = m = mol.AddAtom(ra)
conf.SetAtomPosition(m, (a.x, a.y, a.z))
if a.z:
is_3d = True
if not is_3d:
conf.Set3D(False)
for n, m, b in data.bonds():
mol.AddBond(mapping[n], mapping[m], _bond_map[b.order])
mol.AddConformer(conf)
SanitizeMol(mol)
return mol
|
[
"def",
"to_rdkit_molecule",
"(",
"data",
")",
":",
"mol",
"=",
"RWMol",
"(",
")",
"conf",
"=",
"Conformer",
"(",
")",
"mapping",
"=",
"{",
"}",
"is_3d",
"=",
"False",
"for",
"n",
",",
"a",
"in",
"data",
".",
"atoms",
"(",
")",
":",
"ra",
"=",
"Atom",
"(",
"a",
".",
"number",
")",
"ra",
".",
"SetAtomMapNum",
"(",
"n",
")",
"if",
"a",
".",
"charge",
":",
"ra",
".",
"SetFormalCharge",
"(",
"a",
".",
"charge",
")",
"if",
"a",
".",
"isotope",
"!=",
"a",
".",
"common_isotope",
":",
"ra",
".",
"SetIsotope",
"(",
"a",
".",
"isotope",
")",
"if",
"a",
".",
"radical",
":",
"ra",
".",
"SetNumRadicalElectrons",
"(",
"a",
".",
"radical",
")",
"mapping",
"[",
"n",
"]",
"=",
"m",
"=",
"mol",
".",
"AddAtom",
"(",
"ra",
")",
"conf",
".",
"SetAtomPosition",
"(",
"m",
",",
"(",
"a",
".",
"x",
",",
"a",
".",
"y",
",",
"a",
".",
"z",
")",
")",
"if",
"a",
".",
"z",
":",
"is_3d",
"=",
"True",
"if",
"not",
"is_3d",
":",
"conf",
".",
"Set3D",
"(",
"False",
")",
"for",
"n",
",",
"m",
",",
"b",
"in",
"data",
".",
"bonds",
"(",
")",
":",
"mol",
".",
"AddBond",
"(",
"mapping",
"[",
"n",
"]",
",",
"mapping",
"[",
"m",
"]",
",",
"_bond_map",
"[",
"b",
".",
"order",
"]",
")",
"mol",
".",
"AddConformer",
"(",
"conf",
")",
"SanitizeMol",
"(",
"mol",
")",
"return",
"mol"
] |
MoleculeContainer to RDKit molecule object converter
|
[
"MoleculeContainer",
"to",
"RDKit",
"molecule",
"object",
"converter"
] |
train
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/utils/rdkit.py#L59-L88
|
cimm-kzn/CGRtools
|
CGRtools/algorithms/strings.py
|
StringCommon.__dfs
|
def __dfs(self, start, weights, depth_limit):
"""
modified NX dfs
"""
adj = self._adj
stack = [(start, depth_limit, iter(sorted(adj[start], key=weights)))]
visited = {start}
disconnected = defaultdict(list)
edges = defaultdict(list)
while stack:
parent, depth_now, children = stack[-1]
try:
child = next(children)
except StopIteration:
stack.pop()
else:
if child not in visited:
edges[parent].append(child)
visited.add(child)
if depth_now > 1:
front = adj[child].keys() - {parent}
if front:
stack.append((child, depth_now - 1, iter(sorted(front, key=weights))))
elif child not in disconnected:
disconnected[parent].append(child)
return visited, edges, disconnected
|
python
|
def __dfs(self, start, weights, depth_limit):
"""
modified NX dfs
"""
adj = self._adj
stack = [(start, depth_limit, iter(sorted(adj[start], key=weights)))]
visited = {start}
disconnected = defaultdict(list)
edges = defaultdict(list)
while stack:
parent, depth_now, children = stack[-1]
try:
child = next(children)
except StopIteration:
stack.pop()
else:
if child not in visited:
edges[parent].append(child)
visited.add(child)
if depth_now > 1:
front = adj[child].keys() - {parent}
if front:
stack.append((child, depth_now - 1, iter(sorted(front, key=weights))))
elif child not in disconnected:
disconnected[parent].append(child)
return visited, edges, disconnected
|
[
"def",
"__dfs",
"(",
"self",
",",
"start",
",",
"weights",
",",
"depth_limit",
")",
":",
"adj",
"=",
"self",
".",
"_adj",
"stack",
"=",
"[",
"(",
"start",
",",
"depth_limit",
",",
"iter",
"(",
"sorted",
"(",
"adj",
"[",
"start",
"]",
",",
"key",
"=",
"weights",
")",
")",
")",
"]",
"visited",
"=",
"{",
"start",
"}",
"disconnected",
"=",
"defaultdict",
"(",
"list",
")",
"edges",
"=",
"defaultdict",
"(",
"list",
")",
"while",
"stack",
":",
"parent",
",",
"depth_now",
",",
"children",
"=",
"stack",
"[",
"-",
"1",
"]",
"try",
":",
"child",
"=",
"next",
"(",
"children",
")",
"except",
"StopIteration",
":",
"stack",
".",
"pop",
"(",
")",
"else",
":",
"if",
"child",
"not",
"in",
"visited",
":",
"edges",
"[",
"parent",
"]",
".",
"append",
"(",
"child",
")",
"visited",
".",
"add",
"(",
"child",
")",
"if",
"depth_now",
">",
"1",
":",
"front",
"=",
"adj",
"[",
"child",
"]",
".",
"keys",
"(",
")",
"-",
"{",
"parent",
"}",
"if",
"front",
":",
"stack",
".",
"append",
"(",
"(",
"child",
",",
"depth_now",
"-",
"1",
",",
"iter",
"(",
"sorted",
"(",
"front",
",",
"key",
"=",
"weights",
")",
")",
")",
")",
"elif",
"child",
"not",
"in",
"disconnected",
":",
"disconnected",
"[",
"parent",
"]",
".",
"append",
"(",
"child",
")",
"return",
"visited",
",",
"edges",
",",
"disconnected"
] |
modified NX dfs
|
[
"modified",
"NX",
"dfs"
] |
train
|
https://github.com/cimm-kzn/CGRtools/blob/15a19b04f6e4e1d0dab8e0d32a0877c7f7d70f34/CGRtools/algorithms/strings.py#L130-L158
|
camptocamp/marabunta
|
marabunta/config.py
|
get_args_parser
|
def get_args_parser():
"""Return a parser for command line options."""
parser = argparse.ArgumentParser(
description='Marabunta: Migrating ants for Odoo')
parser.add_argument('--migration-file', '-f',
action=EnvDefault,
envvar='MARABUNTA_MIGRATION_FILE',
required=True,
help='The yaml file containing the migration steps')
parser.add_argument('--database', '-d',
action=EnvDefault,
envvar='MARABUNTA_DATABASE',
required=True,
help="Odoo's database")
parser.add_argument('--db-user', '-u',
action=EnvDefault,
envvar='MARABUNTA_DB_USER',
required=True,
help="Odoo's database user")
parser.add_argument('--db-password', '-w',
action=EnvDefault,
envvar='MARABUNTA_DB_PASSWORD',
required=True,
help="Odoo's database password")
parser.add_argument('--db-port', '-p',
default=os.environ.get('MARABUNTA_DB_PORT', 5432),
help="Odoo's database port")
parser.add_argument('--db-host', '-H',
default=os.environ.get('MARABUNTA_DB_HOST',
'localhost'),
help="Odoo's database host")
parser.add_argument('--mode',
action=EnvDefault,
envvar='MARABUNTA_MODE',
required=False,
help="Specify the mode in which we run the migration,"
"such as 'demo' or 'prod'. Additional operations "
"of this mode will be executed after the main "
"operations and the addons list of this mode "
"will be merged with the main addons list.")
parser.add_argument('--allow-serie',
action=BoolEnvDefault,
required=False,
envvar='MARABUNTA_ALLOW_SERIE',
help='Allow to run more than 1 version upgrade at a '
'time.')
parser.add_argument('--force-version',
required=False,
default=os.environ.get('MARABUNTA_FORCE_VERSION'),
help='Force upgrade of a version, even if it has '
'already been applied.')
group = parser.add_argument_group(
title='Web',
description='Configuration related to the internal web server, '
'used to publish a maintenance page during the migration.',
)
group.add_argument('--web-host',
required=False,
default=os.environ.get('MARABUNTA_WEB_HOST', '0.0.0.0'),
help='Host for the web server')
group.add_argument('--web-port',
required=False,
default=os.environ.get('MARABUNTA_WEB_PORT', 8069),
help='Port for the web server')
group.add_argument('--web-custom-html',
required=False,
default=os.environ.get(
'MARABUNTA_WEB_CUSTOM_HTML'
),
help='Path to a custom html file to publish')
return parser
|
python
|
def get_args_parser():
"""Return a parser for command line options."""
parser = argparse.ArgumentParser(
description='Marabunta: Migrating ants for Odoo')
parser.add_argument('--migration-file', '-f',
action=EnvDefault,
envvar='MARABUNTA_MIGRATION_FILE',
required=True,
help='The yaml file containing the migration steps')
parser.add_argument('--database', '-d',
action=EnvDefault,
envvar='MARABUNTA_DATABASE',
required=True,
help="Odoo's database")
parser.add_argument('--db-user', '-u',
action=EnvDefault,
envvar='MARABUNTA_DB_USER',
required=True,
help="Odoo's database user")
parser.add_argument('--db-password', '-w',
action=EnvDefault,
envvar='MARABUNTA_DB_PASSWORD',
required=True,
help="Odoo's database password")
parser.add_argument('--db-port', '-p',
default=os.environ.get('MARABUNTA_DB_PORT', 5432),
help="Odoo's database port")
parser.add_argument('--db-host', '-H',
default=os.environ.get('MARABUNTA_DB_HOST',
'localhost'),
help="Odoo's database host")
parser.add_argument('--mode',
action=EnvDefault,
envvar='MARABUNTA_MODE',
required=False,
help="Specify the mode in which we run the migration,"
"such as 'demo' or 'prod'. Additional operations "
"of this mode will be executed after the main "
"operations and the addons list of this mode "
"will be merged with the main addons list.")
parser.add_argument('--allow-serie',
action=BoolEnvDefault,
required=False,
envvar='MARABUNTA_ALLOW_SERIE',
help='Allow to run more than 1 version upgrade at a '
'time.')
parser.add_argument('--force-version',
required=False,
default=os.environ.get('MARABUNTA_FORCE_VERSION'),
help='Force upgrade of a version, even if it has '
'already been applied.')
group = parser.add_argument_group(
title='Web',
description='Configuration related to the internal web server, '
'used to publish a maintenance page during the migration.',
)
group.add_argument('--web-host',
required=False,
default=os.environ.get('MARABUNTA_WEB_HOST', '0.0.0.0'),
help='Host for the web server')
group.add_argument('--web-port',
required=False,
default=os.environ.get('MARABUNTA_WEB_PORT', 8069),
help='Port for the web server')
group.add_argument('--web-custom-html',
required=False,
default=os.environ.get(
'MARABUNTA_WEB_CUSTOM_HTML'
),
help='Path to a custom html file to publish')
return parser
|
[
"def",
"get_args_parser",
"(",
")",
":",
"parser",
"=",
"argparse",
".",
"ArgumentParser",
"(",
"description",
"=",
"'Marabunta: Migrating ants for Odoo'",
")",
"parser",
".",
"add_argument",
"(",
"'--migration-file'",
",",
"'-f'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_MIGRATION_FILE'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"'The yaml file containing the migration steps'",
")",
"parser",
".",
"add_argument",
"(",
"'--database'",
",",
"'-d'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_DATABASE'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"\"Odoo's database\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-user'",
",",
"'-u'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_DB_USER'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"\"Odoo's database user\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-password'",
",",
"'-w'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_DB_PASSWORD'",
",",
"required",
"=",
"True",
",",
"help",
"=",
"\"Odoo's database password\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-port'",
",",
"'-p'",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_DB_PORT'",
",",
"5432",
")",
",",
"help",
"=",
"\"Odoo's database port\"",
")",
"parser",
".",
"add_argument",
"(",
"'--db-host'",
",",
"'-H'",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_DB_HOST'",
",",
"'localhost'",
")",
",",
"help",
"=",
"\"Odoo's database host\"",
")",
"parser",
".",
"add_argument",
"(",
"'--mode'",
",",
"action",
"=",
"EnvDefault",
",",
"envvar",
"=",
"'MARABUNTA_MODE'",
",",
"required",
"=",
"False",
",",
"help",
"=",
"\"Specify the mode in which we run the migration,\"",
"\"such as 'demo' or 'prod'. Additional operations \"",
"\"of this mode will be executed after the main \"",
"\"operations and the addons list of this mode \"",
"\"will be merged with the main addons list.\"",
")",
"parser",
".",
"add_argument",
"(",
"'--allow-serie'",
",",
"action",
"=",
"BoolEnvDefault",
",",
"required",
"=",
"False",
",",
"envvar",
"=",
"'MARABUNTA_ALLOW_SERIE'",
",",
"help",
"=",
"'Allow to run more than 1 version upgrade at a '",
"'time.'",
")",
"parser",
".",
"add_argument",
"(",
"'--force-version'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_FORCE_VERSION'",
")",
",",
"help",
"=",
"'Force upgrade of a version, even if it has '",
"'already been applied.'",
")",
"group",
"=",
"parser",
".",
"add_argument_group",
"(",
"title",
"=",
"'Web'",
",",
"description",
"=",
"'Configuration related to the internal web server, '",
"'used to publish a maintenance page during the migration.'",
",",
")",
"group",
".",
"add_argument",
"(",
"'--web-host'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_WEB_HOST'",
",",
"'0.0.0.0'",
")",
",",
"help",
"=",
"'Host for the web server'",
")",
"group",
".",
"add_argument",
"(",
"'--web-port'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_WEB_PORT'",
",",
"8069",
")",
",",
"help",
"=",
"'Port for the web server'",
")",
"group",
".",
"add_argument",
"(",
"'--web-custom-html'",
",",
"required",
"=",
"False",
",",
"default",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'MARABUNTA_WEB_CUSTOM_HTML'",
")",
",",
"help",
"=",
"'Path to a custom html file to publish'",
")",
"return",
"parser"
] |
Return a parser for command line options.
|
[
"Return",
"a",
"parser",
"for",
"command",
"line",
"options",
"."
] |
train
|
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/config.py#L90-L161
|
camptocamp/marabunta
|
marabunta/config.py
|
Config.from_parse_args
|
def from_parse_args(cls, args):
"""Constructor from command line args.
:param args: parse command line arguments
:type args: argparse.ArgumentParser
"""
return cls(args.migration_file,
args.database,
db_user=args.db_user,
db_password=args.db_password,
db_port=args.db_port,
db_host=args.db_host,
mode=args.mode,
allow_serie=args.allow_serie,
force_version=args.force_version,
web_host=args.web_host,
web_port=args.web_port,
web_custom_html=args.web_custom_html,
)
|
python
|
def from_parse_args(cls, args):
"""Constructor from command line args.
:param args: parse command line arguments
:type args: argparse.ArgumentParser
"""
return cls(args.migration_file,
args.database,
db_user=args.db_user,
db_password=args.db_password,
db_port=args.db_port,
db_host=args.db_host,
mode=args.mode,
allow_serie=args.allow_serie,
force_version=args.force_version,
web_host=args.web_host,
web_port=args.web_port,
web_custom_html=args.web_custom_html,
)
|
[
"def",
"from_parse_args",
"(",
"cls",
",",
"args",
")",
":",
"return",
"cls",
"(",
"args",
".",
"migration_file",
",",
"args",
".",
"database",
",",
"db_user",
"=",
"args",
".",
"db_user",
",",
"db_password",
"=",
"args",
".",
"db_password",
",",
"db_port",
"=",
"args",
".",
"db_port",
",",
"db_host",
"=",
"args",
".",
"db_host",
",",
"mode",
"=",
"args",
".",
"mode",
",",
"allow_serie",
"=",
"args",
".",
"allow_serie",
",",
"force_version",
"=",
"args",
".",
"force_version",
",",
"web_host",
"=",
"args",
".",
"web_host",
",",
"web_port",
"=",
"args",
".",
"web_port",
",",
"web_custom_html",
"=",
"args",
".",
"web_custom_html",
",",
")"
] |
Constructor from command line args.
:param args: parse command line arguments
:type args: argparse.ArgumentParser
|
[
"Constructor",
"from",
"command",
"line",
"args",
"."
] |
train
|
https://github.com/camptocamp/marabunta/blob/ec3a7a725c7426d6ed642e0a80119b37880eb91e/marabunta/config.py#L40-L60
|
zetaops/zengine
|
zengine/views/base.py
|
BaseView.set_current
|
def set_current(self, current):
"""
Creates some aliases for attributes of ``current``.
Args:
current: :attr:`~zengine.engine.WFCurrent` object.
"""
self.current = current
self.input = current.input
# self.req = current.request
# self.resp = current.response
self.output = current.output
self.cmd = current.task_data['cmd']
if self.cmd and NEXT_CMD_SPLITTER in self.cmd:
self.cmd, self.next_cmd = self.cmd.split(NEXT_CMD_SPLITTER)
else:
self.next_cmd = None
|
python
|
def set_current(self, current):
"""
Creates some aliases for attributes of ``current``.
Args:
current: :attr:`~zengine.engine.WFCurrent` object.
"""
self.current = current
self.input = current.input
# self.req = current.request
# self.resp = current.response
self.output = current.output
self.cmd = current.task_data['cmd']
if self.cmd and NEXT_CMD_SPLITTER in self.cmd:
self.cmd, self.next_cmd = self.cmd.split(NEXT_CMD_SPLITTER)
else:
self.next_cmd = None
|
[
"def",
"set_current",
"(",
"self",
",",
"current",
")",
":",
"self",
".",
"current",
"=",
"current",
"self",
".",
"input",
"=",
"current",
".",
"input",
"# self.req = current.request",
"# self.resp = current.response",
"self",
".",
"output",
"=",
"current",
".",
"output",
"self",
".",
"cmd",
"=",
"current",
".",
"task_data",
"[",
"'cmd'",
"]",
"if",
"self",
".",
"cmd",
"and",
"NEXT_CMD_SPLITTER",
"in",
"self",
".",
"cmd",
":",
"self",
".",
"cmd",
",",
"self",
".",
"next_cmd",
"=",
"self",
".",
"cmd",
".",
"split",
"(",
"NEXT_CMD_SPLITTER",
")",
"else",
":",
"self",
".",
"next_cmd",
"=",
"None"
] |
Creates some aliases for attributes of ``current``.
Args:
current: :attr:`~zengine.engine.WFCurrent` object.
|
[
"Creates",
"some",
"aliases",
"for",
"attributes",
"of",
"current",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/base.py#L35-L52
|
zetaops/zengine
|
zengine/views/base.py
|
BaseView.form_out
|
def form_out(self, _form=None):
"""
Renders form. Applies form modifiers, then writes
result to response payload. If supplied, given form
object instance will be used instead of view's
default ObjectForm.
Args:
_form (:py:attr:`~zengine.forms.json_form.JsonForm`):
Form object to override `self.object_form`
"""
_form = _form or self.object_form
self.output['forms'] = _form.serialize()
self._add_meta_props(_form)
self.output['forms']['grouping'] = _form.Meta.grouping
self.output['forms']['constraints'] = _form.Meta.constraints
self._patch_form(self.output['forms'])
self.set_client_cmd('form')
|
python
|
def form_out(self, _form=None):
"""
Renders form. Applies form modifiers, then writes
result to response payload. If supplied, given form
object instance will be used instead of view's
default ObjectForm.
Args:
_form (:py:attr:`~zengine.forms.json_form.JsonForm`):
Form object to override `self.object_form`
"""
_form = _form or self.object_form
self.output['forms'] = _form.serialize()
self._add_meta_props(_form)
self.output['forms']['grouping'] = _form.Meta.grouping
self.output['forms']['constraints'] = _form.Meta.constraints
self._patch_form(self.output['forms'])
self.set_client_cmd('form')
|
[
"def",
"form_out",
"(",
"self",
",",
"_form",
"=",
"None",
")",
":",
"_form",
"=",
"_form",
"or",
"self",
".",
"object_form",
"self",
".",
"output",
"[",
"'forms'",
"]",
"=",
"_form",
".",
"serialize",
"(",
")",
"self",
".",
"_add_meta_props",
"(",
"_form",
")",
"self",
".",
"output",
"[",
"'forms'",
"]",
"[",
"'grouping'",
"]",
"=",
"_form",
".",
"Meta",
".",
"grouping",
"self",
".",
"output",
"[",
"'forms'",
"]",
"[",
"'constraints'",
"]",
"=",
"_form",
".",
"Meta",
".",
"constraints",
"self",
".",
"_patch_form",
"(",
"self",
".",
"output",
"[",
"'forms'",
"]",
")",
"self",
".",
"set_client_cmd",
"(",
"'form'",
")"
] |
Renders form. Applies form modifiers, then writes
result to response payload. If supplied, given form
object instance will be used instead of view's
default ObjectForm.
Args:
_form (:py:attr:`~zengine.forms.json_form.JsonForm`):
Form object to override `self.object_form`
|
[
"Renders",
"form",
".",
"Applies",
"form",
"modifiers",
"then",
"writes",
"result",
"to",
"response",
"payload",
".",
"If",
"supplied",
"given",
"form",
"object",
"instance",
"will",
"be",
"used",
"instead",
"of",
"view",
"s",
"default",
"ObjectForm",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/base.py#L86-L103
|
zetaops/zengine
|
zengine/views/base.py
|
BaseView.set_client_cmd
|
def set_client_cmd(self, *args):
"""
Adds given cmd(s) to ``self.output['client_cmd']``
Args:
*args: Client commands.
"""
self.client_cmd.update(args)
self.output['client_cmd'] = list(self.client_cmd)
|
python
|
def set_client_cmd(self, *args):
"""
Adds given cmd(s) to ``self.output['client_cmd']``
Args:
*args: Client commands.
"""
self.client_cmd.update(args)
self.output['client_cmd'] = list(self.client_cmd)
|
[
"def",
"set_client_cmd",
"(",
"self",
",",
"*",
"args",
")",
":",
"self",
".",
"client_cmd",
".",
"update",
"(",
"args",
")",
"self",
".",
"output",
"[",
"'client_cmd'",
"]",
"=",
"list",
"(",
"self",
".",
"client_cmd",
")"
] |
Adds given cmd(s) to ``self.output['client_cmd']``
Args:
*args: Client commands.
|
[
"Adds",
"given",
"cmd",
"(",
"s",
")",
"to",
"self",
".",
"output",
"[",
"client_cmd",
"]"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/views/base.py#L117-L125
|
zetaops/zengine
|
zengine/management_commands.py
|
UpdatePermissions.run
|
def run(self):
"""
Creates new permissions.
"""
from pyoko.lib.utils import get_object_from_path
from zengine.config import settings
model = get_object_from_path(settings.PERMISSION_MODEL)
perm_provider = get_object_from_path(settings.PERMISSION_PROVIDER)
existing_perms = []
new_perms = []
for code, name, desc in perm_provider():
code = six.text_type(code)
if self.manager.args.dry:
exists = model.objects.filter(code=code, name=name)
if exists:
perm = exists[0]
new = False
else:
new = True
perm = model(code=code, name=name)
else:
try:
perm = model.objects.get(code)
existing_perms.append(perm)
except ObjectDoesNotExist:
perm = model(description=desc, code=code, name=name)
perm.key = code
perm.save()
new_perms.append(perm)
# perm, new = model.objects.get_or_create({'description': desc}, code=code, name=name)
# if new:
# new_perms.append(perm)
# else:
# existing_perms.append(perm)
report = "\n\n%s permission(s) were found in DB. " % len(existing_perms)
if new_perms:
report += "\n%s new permission record added. " % len(new_perms)
else:
report += 'No new perms added. '
if new_perms:
if not self.manager.args.dry:
SelectBoxCache.flush(model.__name__)
report += 'Total %s perms exists.' % (len(existing_perms) + len(new_perms))
report = "\n + " + "\n + ".join([p.name or p.code for p in new_perms]) + report
if self.manager.args.dry:
print("\n~~~~~~~~~~~~~~ DRY RUN ~~~~~~~~~~~~~~\n")
print(report + "\n")
|
python
|
def run(self):
"""
Creates new permissions.
"""
from pyoko.lib.utils import get_object_from_path
from zengine.config import settings
model = get_object_from_path(settings.PERMISSION_MODEL)
perm_provider = get_object_from_path(settings.PERMISSION_PROVIDER)
existing_perms = []
new_perms = []
for code, name, desc in perm_provider():
code = six.text_type(code)
if self.manager.args.dry:
exists = model.objects.filter(code=code, name=name)
if exists:
perm = exists[0]
new = False
else:
new = True
perm = model(code=code, name=name)
else:
try:
perm = model.objects.get(code)
existing_perms.append(perm)
except ObjectDoesNotExist:
perm = model(description=desc, code=code, name=name)
perm.key = code
perm.save()
new_perms.append(perm)
# perm, new = model.objects.get_or_create({'description': desc}, code=code, name=name)
# if new:
# new_perms.append(perm)
# else:
# existing_perms.append(perm)
report = "\n\n%s permission(s) were found in DB. " % len(existing_perms)
if new_perms:
report += "\n%s new permission record added. " % len(new_perms)
else:
report += 'No new perms added. '
if new_perms:
if not self.manager.args.dry:
SelectBoxCache.flush(model.__name__)
report += 'Total %s perms exists.' % (len(existing_perms) + len(new_perms))
report = "\n + " + "\n + ".join([p.name or p.code for p in new_perms]) + report
if self.manager.args.dry:
print("\n~~~~~~~~~~~~~~ DRY RUN ~~~~~~~~~~~~~~\n")
print(report + "\n")
|
[
"def",
"run",
"(",
"self",
")",
":",
"from",
"pyoko",
".",
"lib",
".",
"utils",
"import",
"get_object_from_path",
"from",
"zengine",
".",
"config",
"import",
"settings",
"model",
"=",
"get_object_from_path",
"(",
"settings",
".",
"PERMISSION_MODEL",
")",
"perm_provider",
"=",
"get_object_from_path",
"(",
"settings",
".",
"PERMISSION_PROVIDER",
")",
"existing_perms",
"=",
"[",
"]",
"new_perms",
"=",
"[",
"]",
"for",
"code",
",",
"name",
",",
"desc",
"in",
"perm_provider",
"(",
")",
":",
"code",
"=",
"six",
".",
"text_type",
"(",
"code",
")",
"if",
"self",
".",
"manager",
".",
"args",
".",
"dry",
":",
"exists",
"=",
"model",
".",
"objects",
".",
"filter",
"(",
"code",
"=",
"code",
",",
"name",
"=",
"name",
")",
"if",
"exists",
":",
"perm",
"=",
"exists",
"[",
"0",
"]",
"new",
"=",
"False",
"else",
":",
"new",
"=",
"True",
"perm",
"=",
"model",
"(",
"code",
"=",
"code",
",",
"name",
"=",
"name",
")",
"else",
":",
"try",
":",
"perm",
"=",
"model",
".",
"objects",
".",
"get",
"(",
"code",
")",
"existing_perms",
".",
"append",
"(",
"perm",
")",
"except",
"ObjectDoesNotExist",
":",
"perm",
"=",
"model",
"(",
"description",
"=",
"desc",
",",
"code",
"=",
"code",
",",
"name",
"=",
"name",
")",
"perm",
".",
"key",
"=",
"code",
"perm",
".",
"save",
"(",
")",
"new_perms",
".",
"append",
"(",
"perm",
")",
"# perm, new = model.objects.get_or_create({'description': desc}, code=code, name=name)",
"# if new:",
"# new_perms.append(perm)",
"# else:",
"# existing_perms.append(perm)",
"report",
"=",
"\"\\n\\n%s permission(s) were found in DB. \"",
"%",
"len",
"(",
"existing_perms",
")",
"if",
"new_perms",
":",
"report",
"+=",
"\"\\n%s new permission record added. \"",
"%",
"len",
"(",
"new_perms",
")",
"else",
":",
"report",
"+=",
"'No new perms added. '",
"if",
"new_perms",
":",
"if",
"not",
"self",
".",
"manager",
".",
"args",
".",
"dry",
":",
"SelectBoxCache",
".",
"flush",
"(",
"model",
".",
"__name__",
")",
"report",
"+=",
"'Total %s perms exists.'",
"%",
"(",
"len",
"(",
"existing_perms",
")",
"+",
"len",
"(",
"new_perms",
")",
")",
"report",
"=",
"\"\\n + \"",
"+",
"\"\\n + \"",
".",
"join",
"(",
"[",
"p",
".",
"name",
"or",
"p",
".",
"code",
"for",
"p",
"in",
"new_perms",
"]",
")",
"+",
"report",
"if",
"self",
".",
"manager",
".",
"args",
".",
"dry",
":",
"print",
"(",
"\"\\n~~~~~~~~~~~~~~ DRY RUN ~~~~~~~~~~~~~~\\n\"",
")",
"print",
"(",
"report",
"+",
"\"\\n\"",
")"
] |
Creates new permissions.
|
[
"Creates",
"new",
"permissions",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L44-L92
|
zetaops/zengine
|
zengine/management_commands.py
|
CreateUser.run
|
def run(self):
"""
Creates user, encrypts password.
"""
from zengine.models import User
user = User(username=self.manager.args.username, superuser=self.manager.args.super)
user.set_password(self.manager.args.password)
user.save()
print("New user created with ID: %s" % user.key)
|
python
|
def run(self):
"""
Creates user, encrypts password.
"""
from zengine.models import User
user = User(username=self.manager.args.username, superuser=self.manager.args.super)
user.set_password(self.manager.args.password)
user.save()
print("New user created with ID: %s" % user.key)
|
[
"def",
"run",
"(",
"self",
")",
":",
"from",
"zengine",
".",
"models",
"import",
"User",
"user",
"=",
"User",
"(",
"username",
"=",
"self",
".",
"manager",
".",
"args",
".",
"username",
",",
"superuser",
"=",
"self",
".",
"manager",
".",
"args",
".",
"super",
")",
"user",
".",
"set_password",
"(",
"self",
".",
"manager",
".",
"args",
".",
"password",
")",
"user",
".",
"save",
"(",
")",
"print",
"(",
"\"New user created with ID: %s\"",
"%",
"user",
".",
"key",
")"
] |
Creates user, encrypts password.
|
[
"Creates",
"user",
"encrypts",
"password",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L110-L118
|
zetaops/zengine
|
zengine/management_commands.py
|
RunServer.run
|
def run(self):
"""
Starts a development server for the zengine application
"""
print("Development server started on http://%s:%s. \n\nPress Ctrl+C to stop\n" % (
self.manager.args.addr,
self.manager.args.port)
)
if self.manager.args.server_type == 'falcon':
self.run_with_falcon()
elif self.manager.args.server_type == 'tornado':
self.run_with_tornado()
|
python
|
def run(self):
"""
Starts a development server for the zengine application
"""
print("Development server started on http://%s:%s. \n\nPress Ctrl+C to stop\n" % (
self.manager.args.addr,
self.manager.args.port)
)
if self.manager.args.server_type == 'falcon':
self.run_with_falcon()
elif self.manager.args.server_type == 'tornado':
self.run_with_tornado()
|
[
"def",
"run",
"(",
"self",
")",
":",
"print",
"(",
"\"Development server started on http://%s:%s. \\n\\nPress Ctrl+C to stop\\n\"",
"%",
"(",
"self",
".",
"manager",
".",
"args",
".",
"addr",
",",
"self",
".",
"manager",
".",
"args",
".",
"port",
")",
")",
"if",
"self",
".",
"manager",
".",
"args",
".",
"server_type",
"==",
"'falcon'",
":",
"self",
".",
"run_with_falcon",
"(",
")",
"elif",
"self",
".",
"manager",
".",
"args",
".",
"server_type",
"==",
"'tornado'",
":",
"self",
".",
"run_with_tornado",
"(",
")"
] |
Starts a development server for the zengine application
|
[
"Starts",
"a",
"development",
"server",
"for",
"the",
"zengine",
"application"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L139-L150
|
zetaops/zengine
|
zengine/management_commands.py
|
RunServer.run_with_tornado
|
def run_with_tornado(self):
"""
runs the tornado/websockets based test server
"""
from zengine.tornado_server.server import runserver
runserver(self.manager.args.addr, int(self.manager.args.port))
|
python
|
def run_with_tornado(self):
"""
runs the tornado/websockets based test server
"""
from zengine.tornado_server.server import runserver
runserver(self.manager.args.addr, int(self.manager.args.port))
|
[
"def",
"run_with_tornado",
"(",
"self",
")",
":",
"from",
"zengine",
".",
"tornado_server",
".",
"server",
"import",
"runserver",
"runserver",
"(",
"self",
".",
"manager",
".",
"args",
".",
"addr",
",",
"int",
"(",
"self",
".",
"manager",
".",
"args",
".",
"port",
")",
")"
] |
runs the tornado/websockets based test server
|
[
"runs",
"the",
"tornado",
"/",
"websockets",
"based",
"test",
"server"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L152-L157
|
zetaops/zengine
|
zengine/management_commands.py
|
RunServer.run_with_falcon
|
def run_with_falcon(self):
"""
runs the falcon/http based test server
"""
from wsgiref import simple_server
from zengine.server import app
httpd = simple_server.make_server(self.manager.args.addr, int(self.manager.args.port), app)
httpd.serve_forever()
|
python
|
def run_with_falcon(self):
"""
runs the falcon/http based test server
"""
from wsgiref import simple_server
from zengine.server import app
httpd = simple_server.make_server(self.manager.args.addr, int(self.manager.args.port), app)
httpd.serve_forever()
|
[
"def",
"run_with_falcon",
"(",
"self",
")",
":",
"from",
"wsgiref",
"import",
"simple_server",
"from",
"zengine",
".",
"server",
"import",
"app",
"httpd",
"=",
"simple_server",
".",
"make_server",
"(",
"self",
".",
"manager",
".",
"args",
".",
"addr",
",",
"int",
"(",
"self",
".",
"manager",
".",
"args",
".",
"port",
")",
",",
"app",
")",
"httpd",
".",
"serve_forever",
"(",
")"
] |
runs the falcon/http based test server
|
[
"runs",
"the",
"falcon",
"/",
"http",
"based",
"test",
"server"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L159-L166
|
zetaops/zengine
|
zengine/management_commands.py
|
RunWorker.run
|
def run(self):
"""
Starts a development server for the zengine application
"""
from zengine.wf_daemon import run_workers, Worker
worker_count = int(self.manager.args.workers or 1)
if not self.manager.args.daemonize:
print("Starting worker(s)")
if worker_count > 1 or self.manager.args.autoreload:
run_workers(worker_count,
self.manager.args.paths.split(' '),
self.manager.args.daemonize)
else:
worker = Worker()
worker.run()
|
python
|
def run(self):
"""
Starts a development server for the zengine application
"""
from zengine.wf_daemon import run_workers, Worker
worker_count = int(self.manager.args.workers or 1)
if not self.manager.args.daemonize:
print("Starting worker(s)")
if worker_count > 1 or self.manager.args.autoreload:
run_workers(worker_count,
self.manager.args.paths.split(' '),
self.manager.args.daemonize)
else:
worker = Worker()
worker.run()
|
[
"def",
"run",
"(",
"self",
")",
":",
"from",
"zengine",
".",
"wf_daemon",
"import",
"run_workers",
",",
"Worker",
"worker_count",
"=",
"int",
"(",
"self",
".",
"manager",
".",
"args",
".",
"workers",
"or",
"1",
")",
"if",
"not",
"self",
".",
"manager",
".",
"args",
".",
"daemonize",
":",
"print",
"(",
"\"Starting worker(s)\"",
")",
"if",
"worker_count",
">",
"1",
"or",
"self",
".",
"manager",
".",
"args",
".",
"autoreload",
":",
"run_workers",
"(",
"worker_count",
",",
"self",
".",
"manager",
".",
"args",
".",
"paths",
".",
"split",
"(",
"' '",
")",
",",
"self",
".",
"manager",
".",
"args",
".",
"daemonize",
")",
"else",
":",
"worker",
"=",
"Worker",
"(",
")",
"worker",
".",
"run",
"(",
")"
] |
Starts a development server for the zengine application
|
[
"Starts",
"a",
"development",
"server",
"for",
"the",
"zengine",
"application"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L187-L203
|
zetaops/zengine
|
zengine/management_commands.py
|
ExtractTranslations._prepare_domain
|
def _prepare_domain(mapping):
"""Prepare a helper dictionary for the domain to temporarily hold some information."""
# Parse the domain-directory mapping
try:
domain, dir = mapping.split(':')
except ValueError:
print("Please provide the sources in the form of '<domain>:<directory>'")
sys.exit(1)
try:
default_language = settings.TRANSLATION_DOMAINS[domain]
except KeyError:
print("Unknown domain {domain}, check the settings file to make sure"
" this domain is set in TRANSLATION_DOMAINS".format(domain=domain))
sys.exit(1)
# Create a temporary file to hold the `.pot` file for this domain
handle, path = tempfile.mkstemp(prefix='zengine_i18n_', suffix='.pot')
return (domain, {
'default': default_language,
'pot': path,
'source': dir,
})
|
python
|
def _prepare_domain(mapping):
"""Prepare a helper dictionary for the domain to temporarily hold some information."""
# Parse the domain-directory mapping
try:
domain, dir = mapping.split(':')
except ValueError:
print("Please provide the sources in the form of '<domain>:<directory>'")
sys.exit(1)
try:
default_language = settings.TRANSLATION_DOMAINS[domain]
except KeyError:
print("Unknown domain {domain}, check the settings file to make sure"
" this domain is set in TRANSLATION_DOMAINS".format(domain=domain))
sys.exit(1)
# Create a temporary file to hold the `.pot` file for this domain
handle, path = tempfile.mkstemp(prefix='zengine_i18n_', suffix='.pot')
return (domain, {
'default': default_language,
'pot': path,
'source': dir,
})
|
[
"def",
"_prepare_domain",
"(",
"mapping",
")",
":",
"# Parse the domain-directory mapping",
"try",
":",
"domain",
",",
"dir",
"=",
"mapping",
".",
"split",
"(",
"':'",
")",
"except",
"ValueError",
":",
"print",
"(",
"\"Please provide the sources in the form of '<domain>:<directory>'\"",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"try",
":",
"default_language",
"=",
"settings",
".",
"TRANSLATION_DOMAINS",
"[",
"domain",
"]",
"except",
"KeyError",
":",
"print",
"(",
"\"Unknown domain {domain}, check the settings file to make sure\"",
"\" this domain is set in TRANSLATION_DOMAINS\"",
".",
"format",
"(",
"domain",
"=",
"domain",
")",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"# Create a temporary file to hold the `.pot` file for this domain",
"handle",
",",
"path",
"=",
"tempfile",
".",
"mkstemp",
"(",
"prefix",
"=",
"'zengine_i18n_'",
",",
"suffix",
"=",
"'.pot'",
")",
"return",
"(",
"domain",
",",
"{",
"'default'",
":",
"default_language",
",",
"'pot'",
":",
"path",
",",
"'source'",
":",
"dir",
",",
"}",
")"
] |
Prepare a helper dictionary for the domain to temporarily hold some information.
|
[
"Prepare",
"a",
"helper",
"dictionary",
"for",
"the",
"domain",
"to",
"temporarily",
"hold",
"some",
"information",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L227-L248
|
zetaops/zengine
|
zengine/management_commands.py
|
ExtractTranslations._validate_domains
|
def _validate_domains(domains):
"""Check that all domains specified in the settings was provided in the options."""
missing = set(settings.TRANSLATION_DOMAINS.keys()) - set(domains.keys())
if missing:
print('The following domains have been set in the configuration, '
'but their sources were not provided, use the `--source` '
'option to specify their sources: {domains}'.format(domains=', '.join(missing)))
sys.exit(1)
|
python
|
def _validate_domains(domains):
"""Check that all domains specified in the settings was provided in the options."""
missing = set(settings.TRANSLATION_DOMAINS.keys()) - set(domains.keys())
if missing:
print('The following domains have been set in the configuration, '
'but their sources were not provided, use the `--source` '
'option to specify their sources: {domains}'.format(domains=', '.join(missing)))
sys.exit(1)
|
[
"def",
"_validate_domains",
"(",
"domains",
")",
":",
"missing",
"=",
"set",
"(",
"settings",
".",
"TRANSLATION_DOMAINS",
".",
"keys",
"(",
")",
")",
"-",
"set",
"(",
"domains",
".",
"keys",
"(",
")",
")",
"if",
"missing",
":",
"print",
"(",
"'The following domains have been set in the configuration, '",
"'but their sources were not provided, use the `--source` '",
"'option to specify their sources: {domains}'",
".",
"format",
"(",
"domains",
"=",
"', '",
".",
"join",
"(",
"missing",
")",
")",
")",
"sys",
".",
"exit",
"(",
"1",
")"
] |
Check that all domains specified in the settings was provided in the options.
|
[
"Check",
"that",
"all",
"domains",
"specified",
"in",
"the",
"settings",
"was",
"provided",
"in",
"the",
"options",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L251-L258
|
zetaops/zengine
|
zengine/management_commands.py
|
ExtractTranslations._extract_translations
|
def _extract_translations(self, domains):
"""Extract the translations into `.pot` files"""
for domain, options in domains.items():
# Create the extractor
extractor = babel_frontend.extract_messages()
extractor.initialize_options()
# The temporary location to write the `.pot` file
extractor.output_file = options['pot']
# Add the comments marked with 'tn:' to the translation file for translators to read. Strip the marker.
extractor.add_comments = ['tn:']
extractor.strip_comments = True
# The directory where the sources for this domain are located
extractor.input_paths = [options['source']]
# Pass the metadata to the translator
extractor.msgid_bugs_address = self.manager.args.contact
extractor.copyright_holder = self.manager.args.copyright
extractor.version = self.manager.args.version
extractor.project = self.manager.args.project
extractor.finalize_options()
# Add keywords for lazy translation functions, based on their non-lazy variants
extractor.keywords.update({
'gettext_lazy': extractor.keywords['gettext'],
'ngettext_lazy': extractor.keywords['ngettext'],
'__': extractor.keywords['gettext'], # double underscore for lazy
})
# Do the extraction
_run_babel_command(extractor)
|
python
|
def _extract_translations(self, domains):
"""Extract the translations into `.pot` files"""
for domain, options in domains.items():
# Create the extractor
extractor = babel_frontend.extract_messages()
extractor.initialize_options()
# The temporary location to write the `.pot` file
extractor.output_file = options['pot']
# Add the comments marked with 'tn:' to the translation file for translators to read. Strip the marker.
extractor.add_comments = ['tn:']
extractor.strip_comments = True
# The directory where the sources for this domain are located
extractor.input_paths = [options['source']]
# Pass the metadata to the translator
extractor.msgid_bugs_address = self.manager.args.contact
extractor.copyright_holder = self.manager.args.copyright
extractor.version = self.manager.args.version
extractor.project = self.manager.args.project
extractor.finalize_options()
# Add keywords for lazy translation functions, based on their non-lazy variants
extractor.keywords.update({
'gettext_lazy': extractor.keywords['gettext'],
'ngettext_lazy': extractor.keywords['ngettext'],
'__': extractor.keywords['gettext'], # double underscore for lazy
})
# Do the extraction
_run_babel_command(extractor)
|
[
"def",
"_extract_translations",
"(",
"self",
",",
"domains",
")",
":",
"for",
"domain",
",",
"options",
"in",
"domains",
".",
"items",
"(",
")",
":",
"# Create the extractor",
"extractor",
"=",
"babel_frontend",
".",
"extract_messages",
"(",
")",
"extractor",
".",
"initialize_options",
"(",
")",
"# The temporary location to write the `.pot` file",
"extractor",
".",
"output_file",
"=",
"options",
"[",
"'pot'",
"]",
"# Add the comments marked with 'tn:' to the translation file for translators to read. Strip the marker.",
"extractor",
".",
"add_comments",
"=",
"[",
"'tn:'",
"]",
"extractor",
".",
"strip_comments",
"=",
"True",
"# The directory where the sources for this domain are located",
"extractor",
".",
"input_paths",
"=",
"[",
"options",
"[",
"'source'",
"]",
"]",
"# Pass the metadata to the translator",
"extractor",
".",
"msgid_bugs_address",
"=",
"self",
".",
"manager",
".",
"args",
".",
"contact",
"extractor",
".",
"copyright_holder",
"=",
"self",
".",
"manager",
".",
"args",
".",
"copyright",
"extractor",
".",
"version",
"=",
"self",
".",
"manager",
".",
"args",
".",
"version",
"extractor",
".",
"project",
"=",
"self",
".",
"manager",
".",
"args",
".",
"project",
"extractor",
".",
"finalize_options",
"(",
")",
"# Add keywords for lazy translation functions, based on their non-lazy variants",
"extractor",
".",
"keywords",
".",
"update",
"(",
"{",
"'gettext_lazy'",
":",
"extractor",
".",
"keywords",
"[",
"'gettext'",
"]",
",",
"'ngettext_lazy'",
":",
"extractor",
".",
"keywords",
"[",
"'ngettext'",
"]",
",",
"'__'",
":",
"extractor",
".",
"keywords",
"[",
"'gettext'",
"]",
",",
"# double underscore for lazy",
"}",
")",
"# Do the extraction",
"_run_babel_command",
"(",
"extractor",
")"
] |
Extract the translations into `.pot` files
|
[
"Extract",
"the",
"translations",
"into",
".",
"pot",
"files"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L260-L286
|
zetaops/zengine
|
zengine/management_commands.py
|
ExtractTranslations._init_update_po_files
|
def _init_update_po_files(self, domains):
"""Update or initialize the `.po` translation files"""
for language in settings.TRANSLATIONS:
for domain, options in domains.items():
if language == options['default']: continue # Default language of the domain doesn't need translations
if os.path.isfile(_po_path(language, domain)):
# If the translation already exists, update it, keeping the parts already translated
self._update_po_file(language, domain, options['pot'])
else:
# The translation doesn't exist, create a new translation file
self._init_po_file(language, domain, options['pot'])
|
python
|
def _init_update_po_files(self, domains):
"""Update or initialize the `.po` translation files"""
for language in settings.TRANSLATIONS:
for domain, options in domains.items():
if language == options['default']: continue # Default language of the domain doesn't need translations
if os.path.isfile(_po_path(language, domain)):
# If the translation already exists, update it, keeping the parts already translated
self._update_po_file(language, domain, options['pot'])
else:
# The translation doesn't exist, create a new translation file
self._init_po_file(language, domain, options['pot'])
|
[
"def",
"_init_update_po_files",
"(",
"self",
",",
"domains",
")",
":",
"for",
"language",
"in",
"settings",
".",
"TRANSLATIONS",
":",
"for",
"domain",
",",
"options",
"in",
"domains",
".",
"items",
"(",
")",
":",
"if",
"language",
"==",
"options",
"[",
"'default'",
"]",
":",
"continue",
"# Default language of the domain doesn't need translations",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"_po_path",
"(",
"language",
",",
"domain",
")",
")",
":",
"# If the translation already exists, update it, keeping the parts already translated",
"self",
".",
"_update_po_file",
"(",
"language",
",",
"domain",
",",
"options",
"[",
"'pot'",
"]",
")",
"else",
":",
"# The translation doesn't exist, create a new translation file",
"self",
".",
"_init_po_file",
"(",
"language",
",",
"domain",
",",
"options",
"[",
"'pot'",
"]",
")"
] |
Update or initialize the `.po` translation files
|
[
"Update",
"or",
"initialize",
"the",
".",
"po",
"translation",
"files"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L288-L298
|
zetaops/zengine
|
zengine/management_commands.py
|
ExtractTranslations._cleanup
|
def _cleanup(self, domains):
"""Remove the temporary '.pot' files that were created for the domains."""
for option in domains.values():
try:
os.remove(option['pot'])
except (IOError, OSError):
# It is not a problem if we can't actually remove the temporary file
pass
|
python
|
def _cleanup(self, domains):
"""Remove the temporary '.pot' files that were created for the domains."""
for option in domains.values():
try:
os.remove(option['pot'])
except (IOError, OSError):
# It is not a problem if we can't actually remove the temporary file
pass
|
[
"def",
"_cleanup",
"(",
"self",
",",
"domains",
")",
":",
"for",
"option",
"in",
"domains",
".",
"values",
"(",
")",
":",
"try",
":",
"os",
".",
"remove",
"(",
"option",
"[",
"'pot'",
"]",
")",
"except",
"(",
"IOError",
",",
"OSError",
")",
":",
"# It is not a problem if we can't actually remove the temporary file",
"pass"
] |
Remove the temporary '.pot' files that were created for the domains.
|
[
"Remove",
"the",
"temporary",
".",
"pot",
"files",
"that",
"were",
"created",
"for",
"the",
"domains",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L312-L319
|
zetaops/zengine
|
zengine/management_commands.py
|
LoadDiagrams.run
|
def run(self):
"""
read workflows, checks if it's updated,
tries to update if there aren't any running instances of that wf
"""
from zengine.lib.cache import WFSpecNames
if self.manager.args.clear:
self._clear_models()
return
if self.manager.args.wf_path:
paths = self.get_wf_from_path(self.manager.args.wf_path)
else:
paths = self.get_workflows()
self.count = 0
self.do_with_submit(self.load_diagram, paths, threads=self.manager.args.threads)
WFSpecNames().refresh()
print("%s BPMN file loaded" % self.count)
|
python
|
def run(self):
"""
read workflows, checks if it's updated,
tries to update if there aren't any running instances of that wf
"""
from zengine.lib.cache import WFSpecNames
if self.manager.args.clear:
self._clear_models()
return
if self.manager.args.wf_path:
paths = self.get_wf_from_path(self.manager.args.wf_path)
else:
paths = self.get_workflows()
self.count = 0
self.do_with_submit(self.load_diagram, paths, threads=self.manager.args.threads)
WFSpecNames().refresh()
print("%s BPMN file loaded" % self.count)
|
[
"def",
"run",
"(",
"self",
")",
":",
"from",
"zengine",
".",
"lib",
".",
"cache",
"import",
"WFSpecNames",
"if",
"self",
".",
"manager",
".",
"args",
".",
"clear",
":",
"self",
".",
"_clear_models",
"(",
")",
"return",
"if",
"self",
".",
"manager",
".",
"args",
".",
"wf_path",
":",
"paths",
"=",
"self",
".",
"get_wf_from_path",
"(",
"self",
".",
"manager",
".",
"args",
".",
"wf_path",
")",
"else",
":",
"paths",
"=",
"self",
".",
"get_workflows",
"(",
")",
"self",
".",
"count",
"=",
"0",
"self",
".",
"do_with_submit",
"(",
"self",
".",
"load_diagram",
",",
"paths",
",",
"threads",
"=",
"self",
".",
"manager",
".",
"args",
".",
"threads",
")",
"WFSpecNames",
"(",
")",
".",
"refresh",
"(",
")",
"print",
"(",
"\"%s BPMN file loaded\"",
"%",
"self",
".",
"count",
")"
] |
read workflows, checks if it's updated,
tries to update if there aren't any running instances of that wf
|
[
"read",
"workflows",
"checks",
"if",
"it",
"s",
"updated",
"tries",
"to",
"update",
"if",
"there",
"aren",
"t",
"any",
"running",
"instances",
"of",
"that",
"wf"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L450-L472
|
zetaops/zengine
|
zengine/management_commands.py
|
LoadDiagrams.get_wf_from_path
|
def get_wf_from_path(self, path):
"""
load xml from given path
Args:
path: diagram path
Returns:
"""
with open(path) as fp:
content = fp.read()
return [(os.path.basename(os.path.splitext(path)[0]), content), ]
|
python
|
def get_wf_from_path(self, path):
"""
load xml from given path
Args:
path: diagram path
Returns:
"""
with open(path) as fp:
content = fp.read()
return [(os.path.basename(os.path.splitext(path)[0]), content), ]
|
[
"def",
"get_wf_from_path",
"(",
"self",
",",
"path",
")",
":",
"with",
"open",
"(",
"path",
")",
"as",
"fp",
":",
"content",
"=",
"fp",
".",
"read",
"(",
")",
"return",
"[",
"(",
"os",
".",
"path",
".",
"basename",
"(",
"os",
".",
"path",
".",
"splitext",
"(",
"path",
")",
"[",
"0",
"]",
")",
",",
"content",
")",
",",
"]"
] |
load xml from given path
Args:
path: diagram path
Returns:
|
[
"load",
"xml",
"from",
"given",
"path",
"Args",
":",
"path",
":",
"diagram",
"path"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L515-L526
|
zetaops/zengine
|
zengine/management_commands.py
|
LoadDiagrams.get_workflows
|
def get_workflows(self):
"""
Scans and loads all wf found under WORKFLOW_PACKAGES_PATHS
Yields: XML content of diagram file
"""
for pth in settings.WORKFLOW_PACKAGES_PATHS:
for f in glob.glob("%s/*.bpmn" % pth):
with open(f) as fp:
yield os.path.basename(os.path.splitext(f)[0]), fp.read()
|
python
|
def get_workflows(self):
"""
Scans and loads all wf found under WORKFLOW_PACKAGES_PATHS
Yields: XML content of diagram file
"""
for pth in settings.WORKFLOW_PACKAGES_PATHS:
for f in glob.glob("%s/*.bpmn" % pth):
with open(f) as fp:
yield os.path.basename(os.path.splitext(f)[0]), fp.read()
|
[
"def",
"get_workflows",
"(",
"self",
")",
":",
"for",
"pth",
"in",
"settings",
".",
"WORKFLOW_PACKAGES_PATHS",
":",
"for",
"f",
"in",
"glob",
".",
"glob",
"(",
"\"%s/*.bpmn\"",
"%",
"pth",
")",
":",
"with",
"open",
"(",
"f",
")",
"as",
"fp",
":",
"yield",
"os",
".",
"path",
".",
"basename",
"(",
"os",
".",
"path",
".",
"splitext",
"(",
"f",
")",
"[",
"0",
"]",
")",
",",
"fp",
".",
"read",
"(",
")"
] |
Scans and loads all wf found under WORKFLOW_PACKAGES_PATHS
Yields: XML content of diagram file
|
[
"Scans",
"and",
"loads",
"all",
"wf",
"found",
"under",
"WORKFLOW_PACKAGES_PATHS"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L528-L538
|
zetaops/zengine
|
zengine/management_commands.py
|
CheckList.check_migration_and_solr
|
def check_migration_and_solr(self):
"""
The model or models are checked for migrations that need to be done.
Solr is also checked.
"""
from pyoko.db.schema_update import SchemaUpdater
from socket import error as socket_error
from pyoko.conf import settings
from importlib import import_module
import_module(settings.MODELS_MODULE)
registry = import_module('pyoko.model').model_registry
models = [model for model in registry.get_base_models()]
try:
print(__(u"Checking migration and solr ..."))
updater = SchemaUpdater(models, 1, False)
updater.run(check_only=True)
except socket_error as e:
print(__(u"{0}Error not connected, open redis and rabbitmq{1}").format(CheckList.FAIL,
CheckList.ENDC))
|
python
|
def check_migration_and_solr(self):
"""
The model or models are checked for migrations that need to be done.
Solr is also checked.
"""
from pyoko.db.schema_update import SchemaUpdater
from socket import error as socket_error
from pyoko.conf import settings
from importlib import import_module
import_module(settings.MODELS_MODULE)
registry = import_module('pyoko.model').model_registry
models = [model for model in registry.get_base_models()]
try:
print(__(u"Checking migration and solr ..."))
updater = SchemaUpdater(models, 1, False)
updater.run(check_only=True)
except socket_error as e:
print(__(u"{0}Error not connected, open redis and rabbitmq{1}").format(CheckList.FAIL,
CheckList.ENDC))
|
[
"def",
"check_migration_and_solr",
"(",
"self",
")",
":",
"from",
"pyoko",
".",
"db",
".",
"schema_update",
"import",
"SchemaUpdater",
"from",
"socket",
"import",
"error",
"as",
"socket_error",
"from",
"pyoko",
".",
"conf",
"import",
"settings",
"from",
"importlib",
"import",
"import_module",
"import_module",
"(",
"settings",
".",
"MODELS_MODULE",
")",
"registry",
"=",
"import_module",
"(",
"'pyoko.model'",
")",
".",
"model_registry",
"models",
"=",
"[",
"model",
"for",
"model",
"in",
"registry",
".",
"get_base_models",
"(",
")",
"]",
"try",
":",
"print",
"(",
"__",
"(",
"u\"Checking migration and solr ...\"",
")",
")",
"updater",
"=",
"SchemaUpdater",
"(",
"models",
",",
"1",
",",
"False",
")",
"updater",
".",
"run",
"(",
"check_only",
"=",
"True",
")",
"except",
"socket_error",
"as",
"e",
":",
"print",
"(",
"__",
"(",
"u\"{0}Error not connected, open redis and rabbitmq{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
")"
] |
The model or models are checked for migrations that need to be done.
Solr is also checked.
|
[
"The",
"model",
"or",
"models",
"are",
"checked",
"for",
"migrations",
"that",
"need",
"to",
"be",
"done",
".",
"Solr",
"is",
"also",
"checked",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L563-L583
|
zetaops/zengine
|
zengine/management_commands.py
|
CheckList.check_redis
|
def check_redis():
"""
Redis checks the connection
It displays on the screen whether or not you have a connection.
"""
from pyoko.db.connection import cache
from redis.exceptions import ConnectionError
try:
cache.ping()
print(CheckList.OKGREEN + "{0}Redis is working{1}" + CheckList.ENDC)
except ConnectionError as e:
print(__(u"{0}Redis is not working{1} ").format(CheckList.FAIL,
CheckList.ENDC), e.message)
|
python
|
def check_redis():
"""
Redis checks the connection
It displays on the screen whether or not you have a connection.
"""
from pyoko.db.connection import cache
from redis.exceptions import ConnectionError
try:
cache.ping()
print(CheckList.OKGREEN + "{0}Redis is working{1}" + CheckList.ENDC)
except ConnectionError as e:
print(__(u"{0}Redis is not working{1} ").format(CheckList.FAIL,
CheckList.ENDC), e.message)
|
[
"def",
"check_redis",
"(",
")",
":",
"from",
"pyoko",
".",
"db",
".",
"connection",
"import",
"cache",
"from",
"redis",
".",
"exceptions",
"import",
"ConnectionError",
"try",
":",
"cache",
".",
"ping",
"(",
")",
"print",
"(",
"CheckList",
".",
"OKGREEN",
"+",
"\"{0}Redis is working{1}\"",
"+",
"CheckList",
".",
"ENDC",
")",
"except",
"ConnectionError",
"as",
"e",
":",
"print",
"(",
"__",
"(",
"u\"{0}Redis is not working{1} \"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
",",
"e",
".",
"message",
")"
] |
Redis checks the connection
It displays on the screen whether or not you have a connection.
|
[
"Redis",
"checks",
"the",
"connection",
"It",
"displays",
"on",
"the",
"screen",
"whether",
"or",
"not",
"you",
"have",
"a",
"connection",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L586-L599
|
zetaops/zengine
|
zengine/management_commands.py
|
CheckList.check_riak
|
def check_riak():
"""
Riak checks the connection
It displays on the screen whether or not you have a connection.
"""
from pyoko.db.connection import client
from socket import error as socket_error
try:
if client.ping():
print(__(u"{0}Riak is working{1}").format(CheckList.OKGREEN, CheckList.ENDC))
else:
print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL, CheckList.ENDC))
except socket_error as e:
print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL,
CheckList.ENDC), e.message)
|
python
|
def check_riak():
"""
Riak checks the connection
It displays on the screen whether or not you have a connection.
"""
from pyoko.db.connection import client
from socket import error as socket_error
try:
if client.ping():
print(__(u"{0}Riak is working{1}").format(CheckList.OKGREEN, CheckList.ENDC))
else:
print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL, CheckList.ENDC))
except socket_error as e:
print(__(u"{0}Riak is not working{1}").format(CheckList.FAIL,
CheckList.ENDC), e.message)
|
[
"def",
"check_riak",
"(",
")",
":",
"from",
"pyoko",
".",
"db",
".",
"connection",
"import",
"client",
"from",
"socket",
"import",
"error",
"as",
"socket_error",
"try",
":",
"if",
"client",
".",
"ping",
"(",
")",
":",
"print",
"(",
"__",
"(",
"u\"{0}Riak is working{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"OKGREEN",
",",
"CheckList",
".",
"ENDC",
")",
")",
"else",
":",
"print",
"(",
"__",
"(",
"u\"{0}Riak is not working{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
")",
"except",
"socket_error",
"as",
"e",
":",
"print",
"(",
"__",
"(",
"u\"{0}Riak is not working{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
",",
"e",
".",
"message",
")"
] |
Riak checks the connection
It displays on the screen whether or not you have a connection.
|
[
"Riak",
"checks",
"the",
"connection",
"It",
"displays",
"on",
"the",
"screen",
"whether",
"or",
"not",
"you",
"have",
"a",
"connection",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L602-L617
|
zetaops/zengine
|
zengine/management_commands.py
|
CheckList.check_mq_connection
|
def check_mq_connection(self):
"""
RabbitMQ checks the connection
It displays on the screen whether or not you have a connection.
"""
import pika
from zengine.client_queue import BLOCKING_MQ_PARAMS
from pika.exceptions import ProbableAuthenticationError, ConnectionClosed
try:
connection = pika.BlockingConnection(BLOCKING_MQ_PARAMS)
channel = connection.channel()
if channel.is_open:
print(__(u"{0}RabbitMQ is working{1}").format(CheckList.OKGREEN, CheckList.ENDC))
elif self.channel.is_closed or self.channel.is_closing:
print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC))
except ConnectionClosed as e:
print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC), e)
except ProbableAuthenticationError as e:
print(__(u"{0}RabbitMQ username and password wrong{1}").format(CheckList.FAIL,
CheckList.ENDC))
|
python
|
def check_mq_connection(self):
"""
RabbitMQ checks the connection
It displays on the screen whether or not you have a connection.
"""
import pika
from zengine.client_queue import BLOCKING_MQ_PARAMS
from pika.exceptions import ProbableAuthenticationError, ConnectionClosed
try:
connection = pika.BlockingConnection(BLOCKING_MQ_PARAMS)
channel = connection.channel()
if channel.is_open:
print(__(u"{0}RabbitMQ is working{1}").format(CheckList.OKGREEN, CheckList.ENDC))
elif self.channel.is_closed or self.channel.is_closing:
print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC))
except ConnectionClosed as e:
print(__(u"{0}RabbitMQ is not working!{1}").format(CheckList.FAIL, CheckList.ENDC), e)
except ProbableAuthenticationError as e:
print(__(u"{0}RabbitMQ username and password wrong{1}").format(CheckList.FAIL,
CheckList.ENDC))
|
[
"def",
"check_mq_connection",
"(",
"self",
")",
":",
"import",
"pika",
"from",
"zengine",
".",
"client_queue",
"import",
"BLOCKING_MQ_PARAMS",
"from",
"pika",
".",
"exceptions",
"import",
"ProbableAuthenticationError",
",",
"ConnectionClosed",
"try",
":",
"connection",
"=",
"pika",
".",
"BlockingConnection",
"(",
"BLOCKING_MQ_PARAMS",
")",
"channel",
"=",
"connection",
".",
"channel",
"(",
")",
"if",
"channel",
".",
"is_open",
":",
"print",
"(",
"__",
"(",
"u\"{0}RabbitMQ is working{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"OKGREEN",
",",
"CheckList",
".",
"ENDC",
")",
")",
"elif",
"self",
".",
"channel",
".",
"is_closed",
"or",
"self",
".",
"channel",
".",
"is_closing",
":",
"print",
"(",
"__",
"(",
"u\"{0}RabbitMQ is not working!{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
")",
"except",
"ConnectionClosed",
"as",
"e",
":",
"print",
"(",
"__",
"(",
"u\"{0}RabbitMQ is not working!{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
",",
"e",
")",
"except",
"ProbableAuthenticationError",
"as",
"e",
":",
"print",
"(",
"__",
"(",
"u\"{0}RabbitMQ username and password wrong{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
")"
] |
RabbitMQ checks the connection
It displays on the screen whether or not you have a connection.
|
[
"RabbitMQ",
"checks",
"the",
"connection",
"It",
"displays",
"on",
"the",
"screen",
"whether",
"or",
"not",
"you",
"have",
"a",
"connection",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L619-L639
|
zetaops/zengine
|
zengine/management_commands.py
|
CheckList.check_encoding_and_env
|
def check_encoding_and_env():
"""
It brings the environment variables to the screen.
The user checks to see if they are using the correct variables.
"""
import sys
import os
if sys.getfilesystemencoding() in ['utf-8', 'UTF-8']:
print(__(u"{0}File system encoding correct{1}").format(CheckList.OKGREEN,
CheckList.ENDC))
else:
print(__(u"{0}File system encoding wrong!!{1}").format(CheckList.FAIL,
CheckList.ENDC))
check_env_list = ['RIAK_PROTOCOL', 'RIAK_SERVER', 'RIAK_PORT', 'REDIS_SERVER',
'DEFAULT_BUCKET_TYPE', 'PYOKO_SETTINGS',
'MQ_HOST', 'MQ_PORT', 'MQ_USER', 'MQ_VHOST',
]
env = os.environ
for k, v in env.items():
if k in check_env_list:
print(__(u"{0}{1} : {2}{3}").format(CheckList.BOLD, k, v, CheckList.ENDC))
|
python
|
def check_encoding_and_env():
"""
It brings the environment variables to the screen.
The user checks to see if they are using the correct variables.
"""
import sys
import os
if sys.getfilesystemencoding() in ['utf-8', 'UTF-8']:
print(__(u"{0}File system encoding correct{1}").format(CheckList.OKGREEN,
CheckList.ENDC))
else:
print(__(u"{0}File system encoding wrong!!{1}").format(CheckList.FAIL,
CheckList.ENDC))
check_env_list = ['RIAK_PROTOCOL', 'RIAK_SERVER', 'RIAK_PORT', 'REDIS_SERVER',
'DEFAULT_BUCKET_TYPE', 'PYOKO_SETTINGS',
'MQ_HOST', 'MQ_PORT', 'MQ_USER', 'MQ_VHOST',
]
env = os.environ
for k, v in env.items():
if k in check_env_list:
print(__(u"{0}{1} : {2}{3}").format(CheckList.BOLD, k, v, CheckList.ENDC))
|
[
"def",
"check_encoding_and_env",
"(",
")",
":",
"import",
"sys",
"import",
"os",
"if",
"sys",
".",
"getfilesystemencoding",
"(",
")",
"in",
"[",
"'utf-8'",
",",
"'UTF-8'",
"]",
":",
"print",
"(",
"__",
"(",
"u\"{0}File system encoding correct{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"OKGREEN",
",",
"CheckList",
".",
"ENDC",
")",
")",
"else",
":",
"print",
"(",
"__",
"(",
"u\"{0}File system encoding wrong!!{1}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"FAIL",
",",
"CheckList",
".",
"ENDC",
")",
")",
"check_env_list",
"=",
"[",
"'RIAK_PROTOCOL'",
",",
"'RIAK_SERVER'",
",",
"'RIAK_PORT'",
",",
"'REDIS_SERVER'",
",",
"'DEFAULT_BUCKET_TYPE'",
",",
"'PYOKO_SETTINGS'",
",",
"'MQ_HOST'",
",",
"'MQ_PORT'",
",",
"'MQ_USER'",
",",
"'MQ_VHOST'",
",",
"]",
"env",
"=",
"os",
".",
"environ",
"for",
"k",
",",
"v",
"in",
"env",
".",
"items",
"(",
")",
":",
"if",
"k",
"in",
"check_env_list",
":",
"print",
"(",
"__",
"(",
"u\"{0}{1} : {2}{3}\"",
")",
".",
"format",
"(",
"CheckList",
".",
"BOLD",
",",
"k",
",",
"v",
",",
"CheckList",
".",
"ENDC",
")",
")"
] |
It brings the environment variables to the screen.
The user checks to see if they are using the correct variables.
|
[
"It",
"brings",
"the",
"environment",
"variables",
"to",
"the",
"screen",
".",
"The",
"user",
"checks",
"to",
"see",
"if",
"they",
"are",
"using",
"the",
"correct",
"variables",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/management_commands.py#L642-L662
|
LordDarkula/chess_py
|
chess_py/game/game_state.py
|
no_moves
|
def no_moves(position):
"""
Finds if the game is over.
:type: position: Board
:rtype: bool
"""
return position.no_moves(color.white) \
or position.no_moves(color.black)
|
python
|
def no_moves(position):
"""
Finds if the game is over.
:type: position: Board
:rtype: bool
"""
return position.no_moves(color.white) \
or position.no_moves(color.black)
|
[
"def",
"no_moves",
"(",
"position",
")",
":",
"return",
"position",
".",
"no_moves",
"(",
"color",
".",
"white",
")",
"or",
"position",
".",
"no_moves",
"(",
"color",
".",
"black",
")"
] |
Finds if the game is over.
:type: position: Board
:rtype: bool
|
[
"Finds",
"if",
"the",
"game",
"is",
"over",
"."
] |
train
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game_state.py#L13-L21
|
LordDarkula/chess_py
|
chess_py/game/game_state.py
|
is_checkmate
|
def is_checkmate(position, input_color):
"""
Finds if particular King is checkmated.
:type: position: Board
:type: input_color: Color
:rtype: bool
"""
return position.no_moves(input_color) and \
position.get_king(input_color).in_check(position)
|
python
|
def is_checkmate(position, input_color):
"""
Finds if particular King is checkmated.
:type: position: Board
:type: input_color: Color
:rtype: bool
"""
return position.no_moves(input_color) and \
position.get_king(input_color).in_check(position)
|
[
"def",
"is_checkmate",
"(",
"position",
",",
"input_color",
")",
":",
"return",
"position",
".",
"no_moves",
"(",
"input_color",
")",
"and",
"position",
".",
"get_king",
"(",
"input_color",
")",
".",
"in_check",
"(",
"position",
")"
] |
Finds if particular King is checkmated.
:type: position: Board
:type: input_color: Color
:rtype: bool
|
[
"Finds",
"if",
"particular",
"King",
"is",
"checkmated",
"."
] |
train
|
https://github.com/LordDarkula/chess_py/blob/14bebc2f8c49ae25c59375cc83d0b38d8ff7281d/chess_py/game/game_state.py#L24-L33
|
zetaops/zengine
|
zengine/messaging/views.py
|
_paginate
|
def _paginate(self, current_page, query_set, per_page=10):
"""
Handles pagination of object listings.
Args:
current_page int:
Current page number
query_set (:class:`QuerySet<pyoko:pyoko.db.queryset.QuerySet>`):
Object listing queryset.
per_page int:
Objects per page.
Returns:
QuerySet object, pagination data dict as a tuple
"""
total_objects = query_set.count()
total_pages = int(total_objects / per_page or 1)
# add orphans to last page
current_per_page = per_page + (
total_objects % per_page if current_page == total_pages else 0)
pagination_data = dict(page=current_page,
total_pages=total_pages,
total_objects=total_objects,
per_page=current_per_page)
query_set = query_set.set_params(rows=current_per_page, start=(current_page - 1) * per_page)
return query_set, pagination_data
|
python
|
def _paginate(self, current_page, query_set, per_page=10):
"""
Handles pagination of object listings.
Args:
current_page int:
Current page number
query_set (:class:`QuerySet<pyoko:pyoko.db.queryset.QuerySet>`):
Object listing queryset.
per_page int:
Objects per page.
Returns:
QuerySet object, pagination data dict as a tuple
"""
total_objects = query_set.count()
total_pages = int(total_objects / per_page or 1)
# add orphans to last page
current_per_page = per_page + (
total_objects % per_page if current_page == total_pages else 0)
pagination_data = dict(page=current_page,
total_pages=total_pages,
total_objects=total_objects,
per_page=current_per_page)
query_set = query_set.set_params(rows=current_per_page, start=(current_page - 1) * per_page)
return query_set, pagination_data
|
[
"def",
"_paginate",
"(",
"self",
",",
"current_page",
",",
"query_set",
",",
"per_page",
"=",
"10",
")",
":",
"total_objects",
"=",
"query_set",
".",
"count",
"(",
")",
"total_pages",
"=",
"int",
"(",
"total_objects",
"/",
"per_page",
"or",
"1",
")",
"# add orphans to last page",
"current_per_page",
"=",
"per_page",
"+",
"(",
"total_objects",
"%",
"per_page",
"if",
"current_page",
"==",
"total_pages",
"else",
"0",
")",
"pagination_data",
"=",
"dict",
"(",
"page",
"=",
"current_page",
",",
"total_pages",
"=",
"total_pages",
",",
"total_objects",
"=",
"total_objects",
",",
"per_page",
"=",
"current_per_page",
")",
"query_set",
"=",
"query_set",
".",
"set_params",
"(",
"rows",
"=",
"current_per_page",
",",
"start",
"=",
"(",
"current_page",
"-",
"1",
")",
"*",
"per_page",
")",
"return",
"query_set",
",",
"pagination_data"
] |
Handles pagination of object listings.
Args:
current_page int:
Current page number
query_set (:class:`QuerySet<pyoko:pyoko.db.queryset.QuerySet>`):
Object listing queryset.
per_page int:
Objects per page.
Returns:
QuerySet object, pagination data dict as a tuple
|
[
"Handles",
"pagination",
"of",
"object",
"listings",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L72-L97
|
zetaops/zengine
|
zengine/messaging/views.py
|
create_message
|
def create_message(current):
"""
Creates a message for the given channel.
.. code-block:: python
# request:
{
'view':'_zops_create_message',
'message': {
'channel': key, # of channel
'body': string, # message text.,
'type': int, # zengine.messaging.model.MSG_TYPES,
'attachments': [{
'description': string, # can be blank,
'name': string, # file name with extension,
'content': string, # base64 encoded file content
}]}
# response:
{
'status': 'Created',
'code': 201,
'msg_key': key, # key of the message object,
}
"""
msg = current.input['message']
msg_obj = Channel.add_message(msg['channel'], body=msg['body'], typ=msg['type'],
sender=current.user,
title=msg['title'], receiver=msg['receiver'] or None)
current.output = {
'msg_key': msg_obj.key,
'status': 'Created',
'code': 201
}
if 'attachment' in msg:
for atch in msg['attachments']:
typ = current._dedect_file_type(atch['name'], atch['content'])
Attachment(channel_id=msg['channel'], msg=msg_obj, name=atch['name'],
file=atch['content'], description=atch['description'], typ=typ).save()
|
python
|
def create_message(current):
"""
Creates a message for the given channel.
.. code-block:: python
# request:
{
'view':'_zops_create_message',
'message': {
'channel': key, # of channel
'body': string, # message text.,
'type': int, # zengine.messaging.model.MSG_TYPES,
'attachments': [{
'description': string, # can be blank,
'name': string, # file name with extension,
'content': string, # base64 encoded file content
}]}
# response:
{
'status': 'Created',
'code': 201,
'msg_key': key, # key of the message object,
}
"""
msg = current.input['message']
msg_obj = Channel.add_message(msg['channel'], body=msg['body'], typ=msg['type'],
sender=current.user,
title=msg['title'], receiver=msg['receiver'] or None)
current.output = {
'msg_key': msg_obj.key,
'status': 'Created',
'code': 201
}
if 'attachment' in msg:
for atch in msg['attachments']:
typ = current._dedect_file_type(atch['name'], atch['content'])
Attachment(channel_id=msg['channel'], msg=msg_obj, name=atch['name'],
file=atch['content'], description=atch['description'], typ=typ).save()
|
[
"def",
"create_message",
"(",
"current",
")",
":",
"msg",
"=",
"current",
".",
"input",
"[",
"'message'",
"]",
"msg_obj",
"=",
"Channel",
".",
"add_message",
"(",
"msg",
"[",
"'channel'",
"]",
",",
"body",
"=",
"msg",
"[",
"'body'",
"]",
",",
"typ",
"=",
"msg",
"[",
"'type'",
"]",
",",
"sender",
"=",
"current",
".",
"user",
",",
"title",
"=",
"msg",
"[",
"'title'",
"]",
",",
"receiver",
"=",
"msg",
"[",
"'receiver'",
"]",
"or",
"None",
")",
"current",
".",
"output",
"=",
"{",
"'msg_key'",
":",
"msg_obj",
".",
"key",
",",
"'status'",
":",
"'Created'",
",",
"'code'",
":",
"201",
"}",
"if",
"'attachment'",
"in",
"msg",
":",
"for",
"atch",
"in",
"msg",
"[",
"'attachments'",
"]",
":",
"typ",
"=",
"current",
".",
"_dedect_file_type",
"(",
"atch",
"[",
"'name'",
"]",
",",
"atch",
"[",
"'content'",
"]",
")",
"Attachment",
"(",
"channel_id",
"=",
"msg",
"[",
"'channel'",
"]",
",",
"msg",
"=",
"msg_obj",
",",
"name",
"=",
"atch",
"[",
"'name'",
"]",
",",
"file",
"=",
"atch",
"[",
"'content'",
"]",
",",
"description",
"=",
"atch",
"[",
"'description'",
"]",
",",
"typ",
"=",
"typ",
")",
".",
"save",
"(",
")"
] |
Creates a message for the given channel.
.. code-block:: python
# request:
{
'view':'_zops_create_message',
'message': {
'channel': key, # of channel
'body': string, # message text.,
'type': int, # zengine.messaging.model.MSG_TYPES,
'attachments': [{
'description': string, # can be blank,
'name': string, # file name with extension,
'content': string, # base64 encoded file content
}]}
# response:
{
'status': 'Created',
'code': 201,
'msg_key': key, # key of the message object,
}
|
[
"Creates",
"a",
"message",
"for",
"the",
"given",
"channel",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L100-L139
|
zetaops/zengine
|
zengine/messaging/views.py
|
show_channel
|
def show_channel(current, waited=False):
"""
Initial display of channel content.
Returns channel description, members, no of members, last 20 messages etc.
.. code-block:: python
# request:
{
'view':'_zops_show_channel',
'key': key,
}
# response:
{
'channel_key': key,
'description': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'name': string,
'last_messages': [MSG_DICT]
'status': 'OK',
'code': 200
}
"""
ch = Channel(current).objects.get(current.input['key'])
sbs = ch.get_subscription_for_user(current.user_id)
current.output = {'key': current.input['key'],
'description': ch.description,
'name': sbs.name,
'actions': sbs.get_actions(),
'avatar_url': ch.get_avatar(current.user),
'no_of_members': len(ch.subscriber_set),
'member_list': [{'name': sb.user.full_name,
'is_online': sb.user.is_online(),
'avatar_url': sb.user.get_avatar_url()
} for sb in ch.subscriber_set.objects.all()],
'last_messages': [],
'status': 'OK',
'code': 200
}
for msg in ch.get_last_messages():
current.output['last_messages'].insert(0, msg.serialize(current.user))
|
python
|
def show_channel(current, waited=False):
"""
Initial display of channel content.
Returns channel description, members, no of members, last 20 messages etc.
.. code-block:: python
# request:
{
'view':'_zops_show_channel',
'key': key,
}
# response:
{
'channel_key': key,
'description': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'name': string,
'last_messages': [MSG_DICT]
'status': 'OK',
'code': 200
}
"""
ch = Channel(current).objects.get(current.input['key'])
sbs = ch.get_subscription_for_user(current.user_id)
current.output = {'key': current.input['key'],
'description': ch.description,
'name': sbs.name,
'actions': sbs.get_actions(),
'avatar_url': ch.get_avatar(current.user),
'no_of_members': len(ch.subscriber_set),
'member_list': [{'name': sb.user.full_name,
'is_online': sb.user.is_online(),
'avatar_url': sb.user.get_avatar_url()
} for sb in ch.subscriber_set.objects.all()],
'last_messages': [],
'status': 'OK',
'code': 200
}
for msg in ch.get_last_messages():
current.output['last_messages'].insert(0, msg.serialize(current.user))
|
[
"def",
"show_channel",
"(",
"current",
",",
"waited",
"=",
"False",
")",
":",
"ch",
"=",
"Channel",
"(",
"current",
")",
".",
"objects",
".",
"get",
"(",
"current",
".",
"input",
"[",
"'key'",
"]",
")",
"sbs",
"=",
"ch",
".",
"get_subscription_for_user",
"(",
"current",
".",
"user_id",
")",
"current",
".",
"output",
"=",
"{",
"'key'",
":",
"current",
".",
"input",
"[",
"'key'",
"]",
",",
"'description'",
":",
"ch",
".",
"description",
",",
"'name'",
":",
"sbs",
".",
"name",
",",
"'actions'",
":",
"sbs",
".",
"get_actions",
"(",
")",
",",
"'avatar_url'",
":",
"ch",
".",
"get_avatar",
"(",
"current",
".",
"user",
")",
",",
"'no_of_members'",
":",
"len",
"(",
"ch",
".",
"subscriber_set",
")",
",",
"'member_list'",
":",
"[",
"{",
"'name'",
":",
"sb",
".",
"user",
".",
"full_name",
",",
"'is_online'",
":",
"sb",
".",
"user",
".",
"is_online",
"(",
")",
",",
"'avatar_url'",
":",
"sb",
".",
"user",
".",
"get_avatar_url",
"(",
")",
"}",
"for",
"sb",
"in",
"ch",
".",
"subscriber_set",
".",
"objects",
".",
"all",
"(",
")",
"]",
",",
"'last_messages'",
":",
"[",
"]",
",",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}",
"for",
"msg",
"in",
"ch",
".",
"get_last_messages",
"(",
")",
":",
"current",
".",
"output",
"[",
"'last_messages'",
"]",
".",
"insert",
"(",
"0",
",",
"msg",
".",
"serialize",
"(",
"current",
".",
"user",
")",
")"
] |
Initial display of channel content.
Returns channel description, members, no of members, last 20 messages etc.
.. code-block:: python
# request:
{
'view':'_zops_show_channel',
'key': key,
}
# response:
{
'channel_key': key,
'description': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'name': string,
'last_messages': [MSG_DICT]
'status': 'OK',
'code': 200
}
|
[
"Initial",
"display",
"of",
"channel",
"content",
".",
"Returns",
"channel",
"description",
"members",
"no",
"of",
"members",
"last",
"20",
"messages",
"etc",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L142-L189
|
zetaops/zengine
|
zengine/messaging/views.py
|
channel_history
|
def channel_history(current):
"""
Get old messages for a channel. 20 messages per request
.. code-block:: python
# request:
{
'view':'_zops_channel_history,
'channel_key': key,
'timestamp': datetime, # timestamp data of oldest shown message
}
# response:
{
'messages': [MSG_DICT, ],
'status': 'OK',
'code': 200
}
"""
current.output = {
'status': 'OK',
'code': 201,
'messages': []
}
for msg in list(Message.objects.filter(channel_id=current.input['channel_key'],
updated_at__lte=current.input['timestamp'])[:20]):
current.output['messages'].insert(0, msg.serialize(current.user))
# FIXME: looks like pyoko's __lt is broken
# TODO: convert lte to lt and remove this block, when __lt filter fixed
if current.output['messages']:
current.output['messages'].pop(-1)
|
python
|
def channel_history(current):
"""
Get old messages for a channel. 20 messages per request
.. code-block:: python
# request:
{
'view':'_zops_channel_history,
'channel_key': key,
'timestamp': datetime, # timestamp data of oldest shown message
}
# response:
{
'messages': [MSG_DICT, ],
'status': 'OK',
'code': 200
}
"""
current.output = {
'status': 'OK',
'code': 201,
'messages': []
}
for msg in list(Message.objects.filter(channel_id=current.input['channel_key'],
updated_at__lte=current.input['timestamp'])[:20]):
current.output['messages'].insert(0, msg.serialize(current.user))
# FIXME: looks like pyoko's __lt is broken
# TODO: convert lte to lt and remove this block, when __lt filter fixed
if current.output['messages']:
current.output['messages'].pop(-1)
|
[
"def",
"channel_history",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"201",
",",
"'messages'",
":",
"[",
"]",
"}",
"for",
"msg",
"in",
"list",
"(",
"Message",
".",
"objects",
".",
"filter",
"(",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
",",
"updated_at__lte",
"=",
"current",
".",
"input",
"[",
"'timestamp'",
"]",
")",
"[",
":",
"20",
"]",
")",
":",
"current",
".",
"output",
"[",
"'messages'",
"]",
".",
"insert",
"(",
"0",
",",
"msg",
".",
"serialize",
"(",
"current",
".",
"user",
")",
")",
"# FIXME: looks like pyoko's __lt is broken",
"# TODO: convert lte to lt and remove this block, when __lt filter fixed",
"if",
"current",
".",
"output",
"[",
"'messages'",
"]",
":",
"current",
".",
"output",
"[",
"'messages'",
"]",
".",
"pop",
"(",
"-",
"1",
")"
] |
Get old messages for a channel. 20 messages per request
.. code-block:: python
# request:
{
'view':'_zops_channel_history,
'channel_key': key,
'timestamp': datetime, # timestamp data of oldest shown message
}
# response:
{
'messages': [MSG_DICT, ],
'status': 'OK',
'code': 200
}
|
[
"Get",
"old",
"messages",
"for",
"a",
"channel",
".",
"20",
"messages",
"per",
"request"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L192-L224
|
zetaops/zengine
|
zengine/messaging/views.py
|
report_last_seen_message
|
def report_last_seen_message(current):
"""
Push timestamp of latest message of an ACTIVE channel.
This view should be called with timestamp of latest message;
- When user opens (clicks on) a channel.
- Periodically (eg: setInterval for 15secs) while user staying in a channel.
.. code-block:: python
# request:
{
'view':'_zops_last_seen_msg',
'channel_key': key,
'key': key,
'timestamp': datetime,
}
# response:
{
'status': 'OK',
'code': 200,
}
"""
sbs = Subscriber(current).objects.filter(channel_id=current.input['channel_key'],
user_id=current.user_id)[0]
sbs.last_seen_msg_time = current.input['timestamp']
sbs.save()
current.output = {
'status': 'OK',
'code': 200}
|
python
|
def report_last_seen_message(current):
"""
Push timestamp of latest message of an ACTIVE channel.
This view should be called with timestamp of latest message;
- When user opens (clicks on) a channel.
- Periodically (eg: setInterval for 15secs) while user staying in a channel.
.. code-block:: python
# request:
{
'view':'_zops_last_seen_msg',
'channel_key': key,
'key': key,
'timestamp': datetime,
}
# response:
{
'status': 'OK',
'code': 200,
}
"""
sbs = Subscriber(current).objects.filter(channel_id=current.input['channel_key'],
user_id=current.user_id)[0]
sbs.last_seen_msg_time = current.input['timestamp']
sbs.save()
current.output = {
'status': 'OK',
'code': 200}
|
[
"def",
"report_last_seen_message",
"(",
"current",
")",
":",
"sbs",
"=",
"Subscriber",
"(",
"current",
")",
".",
"objects",
".",
"filter",
"(",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
",",
"user_id",
"=",
"current",
".",
"user_id",
")",
"[",
"0",
"]",
"sbs",
".",
"last_seen_msg_time",
"=",
"current",
".",
"input",
"[",
"'timestamp'",
"]",
"sbs",
".",
"save",
"(",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}"
] |
Push timestamp of latest message of an ACTIVE channel.
This view should be called with timestamp of latest message;
- When user opens (clicks on) a channel.
- Periodically (eg: setInterval for 15secs) while user staying in a channel.
.. code-block:: python
# request:
{
'view':'_zops_last_seen_msg',
'channel_key': key,
'key': key,
'timestamp': datetime,
}
# response:
{
'status': 'OK',
'code': 200,
}
|
[
"Push",
"timestamp",
"of",
"latest",
"message",
"of",
"an",
"ACTIVE",
"channel",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L227-L258
|
zetaops/zengine
|
zengine/messaging/views.py
|
list_channels
|
def list_channels(current):
"""
List channel memberships of current user
.. code-block:: python
# request:
{
'view':'_zops_list_channels',
}
# response:
{
'channels': [
{'name': string, # name of channel
'key': key, # key of channel
'unread': int, # unread message count
'type': int, # channel type,
# 15: public channels (chat room/broadcast channel distinction
comes from "read_only" flag)
# 10: direct channels
# 5: one and only private channel which is "Notifications"
'read_only': boolean,
# true if this is a read-only subscription to a broadcast channel
# false if it's a public chat room
'actions':[('action name', 'view name'),]
},]
}
"""
current.output = {
'status': 'OK',
'code': 200,
'channels': []}
for sbs in current.user.subscriptions.objects.filter(is_visible=True):
try:
current.output['channels'].append(sbs.get_channel_listing())
except ObjectDoesNotExist:
# FIXME: This should not happen,
log.exception("UNPAIRED DIRECT EXCHANGES!!!!")
sbs.delete()
|
python
|
def list_channels(current):
"""
List channel memberships of current user
.. code-block:: python
# request:
{
'view':'_zops_list_channels',
}
# response:
{
'channels': [
{'name': string, # name of channel
'key': key, # key of channel
'unread': int, # unread message count
'type': int, # channel type,
# 15: public channels (chat room/broadcast channel distinction
comes from "read_only" flag)
# 10: direct channels
# 5: one and only private channel which is "Notifications"
'read_only': boolean,
# true if this is a read-only subscription to a broadcast channel
# false if it's a public chat room
'actions':[('action name', 'view name'),]
},]
}
"""
current.output = {
'status': 'OK',
'code': 200,
'channels': []}
for sbs in current.user.subscriptions.objects.filter(is_visible=True):
try:
current.output['channels'].append(sbs.get_channel_listing())
except ObjectDoesNotExist:
# FIXME: This should not happen,
log.exception("UNPAIRED DIRECT EXCHANGES!!!!")
sbs.delete()
|
[
"def",
"list_channels",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
",",
"'channels'",
":",
"[",
"]",
"}",
"for",
"sbs",
"in",
"current",
".",
"user",
".",
"subscriptions",
".",
"objects",
".",
"filter",
"(",
"is_visible",
"=",
"True",
")",
":",
"try",
":",
"current",
".",
"output",
"[",
"'channels'",
"]",
".",
"append",
"(",
"sbs",
".",
"get_channel_listing",
"(",
")",
")",
"except",
"ObjectDoesNotExist",
":",
"# FIXME: This should not happen,",
"log",
".",
"exception",
"(",
"\"UNPAIRED DIRECT EXCHANGES!!!!\"",
")",
"sbs",
".",
"delete",
"(",
")"
] |
List channel memberships of current user
.. code-block:: python
# request:
{
'view':'_zops_list_channels',
}
# response:
{
'channels': [
{'name': string, # name of channel
'key': key, # key of channel
'unread': int, # unread message count
'type': int, # channel type,
# 15: public channels (chat room/broadcast channel distinction
comes from "read_only" flag)
# 10: direct channels
# 5: one and only private channel which is "Notifications"
'read_only': boolean,
# true if this is a read-only subscription to a broadcast channel
# false if it's a public chat room
'actions':[('action name', 'view name'),]
},]
}
|
[
"List",
"channel",
"memberships",
"of",
"current",
"user"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L261-L302
|
zetaops/zengine
|
zengine/messaging/views.py
|
unread_count
|
def unread_count(current):
"""
Number of unread messages for current user
.. code-block:: python
# request:
{
'view':'_zops_unread_count',
}
# response:
{
'status': 'OK',
'code': 200,
'notifications': int,
'messages': int,
}
"""
unread_ntf = 0
unread_msg = 0
for sbs in current.user.subscriptions.objects.filter(is_visible=True):
try:
if sbs.channel.key == current.user.prv_exchange:
unread_ntf += sbs.unread_count()
else:
unread_msg += sbs.unread_count()
except ObjectDoesNotExist:
# FIXME: This should not happen,
log.exception("MULTIPLE PRV EXCHANGES!!!!")
sbs.delete()
current.output = {
'status': 'OK',
'code': 200,
'notifications': unread_ntf,
'messages': unread_msg
}
|
python
|
def unread_count(current):
"""
Number of unread messages for current user
.. code-block:: python
# request:
{
'view':'_zops_unread_count',
}
# response:
{
'status': 'OK',
'code': 200,
'notifications': int,
'messages': int,
}
"""
unread_ntf = 0
unread_msg = 0
for sbs in current.user.subscriptions.objects.filter(is_visible=True):
try:
if sbs.channel.key == current.user.prv_exchange:
unread_ntf += sbs.unread_count()
else:
unread_msg += sbs.unread_count()
except ObjectDoesNotExist:
# FIXME: This should not happen,
log.exception("MULTIPLE PRV EXCHANGES!!!!")
sbs.delete()
current.output = {
'status': 'OK',
'code': 200,
'notifications': unread_ntf,
'messages': unread_msg
}
|
[
"def",
"unread_count",
"(",
"current",
")",
":",
"unread_ntf",
"=",
"0",
"unread_msg",
"=",
"0",
"for",
"sbs",
"in",
"current",
".",
"user",
".",
"subscriptions",
".",
"objects",
".",
"filter",
"(",
"is_visible",
"=",
"True",
")",
":",
"try",
":",
"if",
"sbs",
".",
"channel",
".",
"key",
"==",
"current",
".",
"user",
".",
"prv_exchange",
":",
"unread_ntf",
"+=",
"sbs",
".",
"unread_count",
"(",
")",
"else",
":",
"unread_msg",
"+=",
"sbs",
".",
"unread_count",
"(",
")",
"except",
"ObjectDoesNotExist",
":",
"# FIXME: This should not happen,",
"log",
".",
"exception",
"(",
"\"MULTIPLE PRV EXCHANGES!!!!\"",
")",
"sbs",
".",
"delete",
"(",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
",",
"'notifications'",
":",
"unread_ntf",
",",
"'messages'",
":",
"unread_msg",
"}"
] |
Number of unread messages for current user
.. code-block:: python
# request:
{
'view':'_zops_unread_count',
}
# response:
{
'status': 'OK',
'code': 200,
'notifications': int,
'messages': int,
}
|
[
"Number",
"of",
"unread",
"messages",
"for",
"current",
"user"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L305-L342
|
zetaops/zengine
|
zengine/messaging/views.py
|
get_notifications
|
def get_notifications(current):
"""
Returns last N notifications for current user
.. code-block:: python
# request:
{
'view':'_zops_unread_messages',
'amount': int, # Optional, defaults to 8
}
# response:
{
'status': 'OK',
'code': 200,
'notifications': [{'title':string,
'body': string,
'channel_key': key,
'type': int,
'url': string, # could be a in app JS URL prefixed with "#" or
# full blown URL prefixed with "http"
'message_key': key,
'timestamp': datetime},],
}
"""
current.output = {
'status': 'OK',
'code': 200,
'notifications': [],
}
amount = current.input.get('amount', 8)
try:
notif_sbs = current.user.subscriptions.objects.get(channel_id=current.user.prv_exchange)
except MultipleObjectsReturned:
# FIXME: This should not happen,
log.exception("MULTIPLE PRV EXCHANGES!!!!")
sbs = current.user.subscriptions.objects.filter(channel_id=current.user.prv_exchange)
sbs[0].delete()
notif_sbs = sbs[1]
for msg in notif_sbs.channel.message_set.objects.all()[:amount]:
current.output['notifications'].insert(0, {
'title': msg.msg_title,
'body': msg.body,
'type': msg.typ,
'url': msg.url,
'channel_key': msg.channel.key,
'message_key': msg.key,
'timestamp': msg.updated_at})
|
python
|
def get_notifications(current):
"""
Returns last N notifications for current user
.. code-block:: python
# request:
{
'view':'_zops_unread_messages',
'amount': int, # Optional, defaults to 8
}
# response:
{
'status': 'OK',
'code': 200,
'notifications': [{'title':string,
'body': string,
'channel_key': key,
'type': int,
'url': string, # could be a in app JS URL prefixed with "#" or
# full blown URL prefixed with "http"
'message_key': key,
'timestamp': datetime},],
}
"""
current.output = {
'status': 'OK',
'code': 200,
'notifications': [],
}
amount = current.input.get('amount', 8)
try:
notif_sbs = current.user.subscriptions.objects.get(channel_id=current.user.prv_exchange)
except MultipleObjectsReturned:
# FIXME: This should not happen,
log.exception("MULTIPLE PRV EXCHANGES!!!!")
sbs = current.user.subscriptions.objects.filter(channel_id=current.user.prv_exchange)
sbs[0].delete()
notif_sbs = sbs[1]
for msg in notif_sbs.channel.message_set.objects.all()[:amount]:
current.output['notifications'].insert(0, {
'title': msg.msg_title,
'body': msg.body,
'type': msg.typ,
'url': msg.url,
'channel_key': msg.channel.key,
'message_key': msg.key,
'timestamp': msg.updated_at})
|
[
"def",
"get_notifications",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
",",
"'notifications'",
":",
"[",
"]",
",",
"}",
"amount",
"=",
"current",
".",
"input",
".",
"get",
"(",
"'amount'",
",",
"8",
")",
"try",
":",
"notif_sbs",
"=",
"current",
".",
"user",
".",
"subscriptions",
".",
"objects",
".",
"get",
"(",
"channel_id",
"=",
"current",
".",
"user",
".",
"prv_exchange",
")",
"except",
"MultipleObjectsReturned",
":",
"# FIXME: This should not happen,",
"log",
".",
"exception",
"(",
"\"MULTIPLE PRV EXCHANGES!!!!\"",
")",
"sbs",
"=",
"current",
".",
"user",
".",
"subscriptions",
".",
"objects",
".",
"filter",
"(",
"channel_id",
"=",
"current",
".",
"user",
".",
"prv_exchange",
")",
"sbs",
"[",
"0",
"]",
".",
"delete",
"(",
")",
"notif_sbs",
"=",
"sbs",
"[",
"1",
"]",
"for",
"msg",
"in",
"notif_sbs",
".",
"channel",
".",
"message_set",
".",
"objects",
".",
"all",
"(",
")",
"[",
":",
"amount",
"]",
":",
"current",
".",
"output",
"[",
"'notifications'",
"]",
".",
"insert",
"(",
"0",
",",
"{",
"'title'",
":",
"msg",
".",
"msg_title",
",",
"'body'",
":",
"msg",
".",
"body",
",",
"'type'",
":",
"msg",
".",
"typ",
",",
"'url'",
":",
"msg",
".",
"url",
",",
"'channel_key'",
":",
"msg",
".",
"channel",
".",
"key",
",",
"'message_key'",
":",
"msg",
".",
"key",
",",
"'timestamp'",
":",
"msg",
".",
"updated_at",
"}",
")"
] |
Returns last N notifications for current user
.. code-block:: python
# request:
{
'view':'_zops_unread_messages',
'amount': int, # Optional, defaults to 8
}
# response:
{
'status': 'OK',
'code': 200,
'notifications': [{'title':string,
'body': string,
'channel_key': key,
'type': int,
'url': string, # could be a in app JS URL prefixed with "#" or
# full blown URL prefixed with "http"
'message_key': key,
'timestamp': datetime},],
}
|
[
"Returns",
"last",
"N",
"notifications",
"for",
"current",
"user"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L345-L394
|
zetaops/zengine
|
zengine/messaging/views.py
|
create_channel
|
def create_channel(current):
"""
Create a public channel. Can be a broadcast channel or normal chat room.
Chat room and broadcast distinction will be made at user subscription phase.
.. code-block:: python
# request:
{
'view':'_zops_create_channel',
'name': string,
'description': string,
}
# response:
{
'description': string,
'name': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'last_messages': [MSG_DICT]
'status': 'Created',
'code': 201,
'key': key, # of just created channel
}
"""
channel = Channel(name=current.input['name'],
description=current.input['description'],
owner=current.user,
typ=15).save()
with BlockSave(Subscriber):
Subscriber.objects.get_or_create(user=channel.owner,
channel=channel,
can_manage=True,
can_leave=False)
current.input['key'] = channel.key
show_channel(current)
current.output.update({
'status': 'Created',
'code': 201
})
|
python
|
def create_channel(current):
"""
Create a public channel. Can be a broadcast channel or normal chat room.
Chat room and broadcast distinction will be made at user subscription phase.
.. code-block:: python
# request:
{
'view':'_zops_create_channel',
'name': string,
'description': string,
}
# response:
{
'description': string,
'name': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'last_messages': [MSG_DICT]
'status': 'Created',
'code': 201,
'key': key, # of just created channel
}
"""
channel = Channel(name=current.input['name'],
description=current.input['description'],
owner=current.user,
typ=15).save()
with BlockSave(Subscriber):
Subscriber.objects.get_or_create(user=channel.owner,
channel=channel,
can_manage=True,
can_leave=False)
current.input['key'] = channel.key
show_channel(current)
current.output.update({
'status': 'Created',
'code': 201
})
|
[
"def",
"create_channel",
"(",
"current",
")",
":",
"channel",
"=",
"Channel",
"(",
"name",
"=",
"current",
".",
"input",
"[",
"'name'",
"]",
",",
"description",
"=",
"current",
".",
"input",
"[",
"'description'",
"]",
",",
"owner",
"=",
"current",
".",
"user",
",",
"typ",
"=",
"15",
")",
".",
"save",
"(",
")",
"with",
"BlockSave",
"(",
"Subscriber",
")",
":",
"Subscriber",
".",
"objects",
".",
"get_or_create",
"(",
"user",
"=",
"channel",
".",
"owner",
",",
"channel",
"=",
"channel",
",",
"can_manage",
"=",
"True",
",",
"can_leave",
"=",
"False",
")",
"current",
".",
"input",
"[",
"'key'",
"]",
"=",
"channel",
".",
"key",
"show_channel",
"(",
"current",
")",
"current",
".",
"output",
".",
"update",
"(",
"{",
"'status'",
":",
"'Created'",
",",
"'code'",
":",
"201",
"}",
")"
] |
Create a public channel. Can be a broadcast channel or normal chat room.
Chat room and broadcast distinction will be made at user subscription phase.
.. code-block:: python
# request:
{
'view':'_zops_create_channel',
'name': string,
'description': string,
}
# response:
{
'description': string,
'name': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'last_messages': [MSG_DICT]
'status': 'Created',
'code': 201,
'key': key, # of just created channel
}
|
[
"Create",
"a",
"public",
"channel",
".",
"Can",
"be",
"a",
"broadcast",
"channel",
"or",
"normal",
"chat",
"room",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L397-L442
|
zetaops/zengine
|
zengine/messaging/views.py
|
add_members
|
def add_members(current):
"""
Subscribe member(s) to a channel
.. code-block:: python
# request:
{
'view':'_zops_add_members',
'channel_key': key,
'read_only': boolean, # true if this is a Broadcast channel,
# false if it's a normal chat room
'members': [key, key],
}
# response:
{
'existing': [key,], # existing members
'newly_added': [key,], # newly added members
'status': 'Created',
'code': 201
}
"""
newly_added, existing = [], []
read_only = current.input['read_only']
for member_key in current.input['members']:
sb, new = Subscriber(current).objects.get_or_create(user_id=member_key,
read_only=read_only,
channel_id=current.input['channel_key'])
if new:
newly_added.append(member_key)
else:
existing.append(member_key)
current.output = {
'existing': existing,
'newly_added': newly_added,
'status': 'OK',
'code': 201
}
|
python
|
def add_members(current):
"""
Subscribe member(s) to a channel
.. code-block:: python
# request:
{
'view':'_zops_add_members',
'channel_key': key,
'read_only': boolean, # true if this is a Broadcast channel,
# false if it's a normal chat room
'members': [key, key],
}
# response:
{
'existing': [key,], # existing members
'newly_added': [key,], # newly added members
'status': 'Created',
'code': 201
}
"""
newly_added, existing = [], []
read_only = current.input['read_only']
for member_key in current.input['members']:
sb, new = Subscriber(current).objects.get_or_create(user_id=member_key,
read_only=read_only,
channel_id=current.input['channel_key'])
if new:
newly_added.append(member_key)
else:
existing.append(member_key)
current.output = {
'existing': existing,
'newly_added': newly_added,
'status': 'OK',
'code': 201
}
|
[
"def",
"add_members",
"(",
"current",
")",
":",
"newly_added",
",",
"existing",
"=",
"[",
"]",
",",
"[",
"]",
"read_only",
"=",
"current",
".",
"input",
"[",
"'read_only'",
"]",
"for",
"member_key",
"in",
"current",
".",
"input",
"[",
"'members'",
"]",
":",
"sb",
",",
"new",
"=",
"Subscriber",
"(",
"current",
")",
".",
"objects",
".",
"get_or_create",
"(",
"user_id",
"=",
"member_key",
",",
"read_only",
"=",
"read_only",
",",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
")",
"if",
"new",
":",
"newly_added",
".",
"append",
"(",
"member_key",
")",
"else",
":",
"existing",
".",
"append",
"(",
"member_key",
")",
"current",
".",
"output",
"=",
"{",
"'existing'",
":",
"existing",
",",
"'newly_added'",
":",
"newly_added",
",",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"201",
"}"
] |
Subscribe member(s) to a channel
.. code-block:: python
# request:
{
'view':'_zops_add_members',
'channel_key': key,
'read_only': boolean, # true if this is a Broadcast channel,
# false if it's a normal chat room
'members': [key, key],
}
# response:
{
'existing': [key,], # existing members
'newly_added': [key,], # newly added members
'status': 'Created',
'code': 201
}
|
[
"Subscribe",
"member",
"(",
"s",
")",
"to",
"a",
"channel"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L445-L484
|
zetaops/zengine
|
zengine/messaging/views.py
|
add_unit_to_channel
|
def add_unit_to_channel(current):
"""
Subscribe users of a given unit to given channel
JSON API:
.. code-block:: python
# request:
{
'view':'_zops_add_unit_to_channel',
'unit_key': key,
'channel_key': key,
'read_only': boolean, # true if this is a Broadcast channel,
# false if it's a normal chat room
}
# response:
{
'existing': [key,], # existing members
'newly_added': [key,], # newly added members
'status': 'Created',
'code': 201
}
"""
read_only = current.input['read_only']
newly_added, existing = [], []
for member_key in UnitModel.get_user_keys(current, current.input['unit_key']):
sb, new = Subscriber(current).objects.get_or_create(user_id=member_key,
read_only=read_only,
channel_id=current.input['channel_key'])
if new:
newly_added.append(member_key)
else:
existing.append(member_key)
current.output = {
'existing': existing,
'newly_added': newly_added,
'status': 'OK',
'code': 201
}
|
python
|
def add_unit_to_channel(current):
"""
Subscribe users of a given unit to given channel
JSON API:
.. code-block:: python
# request:
{
'view':'_zops_add_unit_to_channel',
'unit_key': key,
'channel_key': key,
'read_only': boolean, # true if this is a Broadcast channel,
# false if it's a normal chat room
}
# response:
{
'existing': [key,], # existing members
'newly_added': [key,], # newly added members
'status': 'Created',
'code': 201
}
"""
read_only = current.input['read_only']
newly_added, existing = [], []
for member_key in UnitModel.get_user_keys(current, current.input['unit_key']):
sb, new = Subscriber(current).objects.get_or_create(user_id=member_key,
read_only=read_only,
channel_id=current.input['channel_key'])
if new:
newly_added.append(member_key)
else:
existing.append(member_key)
current.output = {
'existing': existing,
'newly_added': newly_added,
'status': 'OK',
'code': 201
}
|
[
"def",
"add_unit_to_channel",
"(",
"current",
")",
":",
"read_only",
"=",
"current",
".",
"input",
"[",
"'read_only'",
"]",
"newly_added",
",",
"existing",
"=",
"[",
"]",
",",
"[",
"]",
"for",
"member_key",
"in",
"UnitModel",
".",
"get_user_keys",
"(",
"current",
",",
"current",
".",
"input",
"[",
"'unit_key'",
"]",
")",
":",
"sb",
",",
"new",
"=",
"Subscriber",
"(",
"current",
")",
".",
"objects",
".",
"get_or_create",
"(",
"user_id",
"=",
"member_key",
",",
"read_only",
"=",
"read_only",
",",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
")",
"if",
"new",
":",
"newly_added",
".",
"append",
"(",
"member_key",
")",
"else",
":",
"existing",
".",
"append",
"(",
"member_key",
")",
"current",
".",
"output",
"=",
"{",
"'existing'",
":",
"existing",
",",
"'newly_added'",
":",
"newly_added",
",",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"201",
"}"
] |
Subscribe users of a given unit to given channel
JSON API:
.. code-block:: python
# request:
{
'view':'_zops_add_unit_to_channel',
'unit_key': key,
'channel_key': key,
'read_only': boolean, # true if this is a Broadcast channel,
# false if it's a normal chat room
}
# response:
{
'existing': [key,], # existing members
'newly_added': [key,], # newly added members
'status': 'Created',
'code': 201
}
|
[
"Subscribe",
"users",
"of",
"a",
"given",
"unit",
"to",
"given",
"channel"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L487-L528
|
zetaops/zengine
|
zengine/messaging/views.py
|
search_user
|
def search_user(current):
"""
Search users for adding to a public room
or creating one to one direct messaging
.. code-block:: python
# request:
{
'view':'_zops_search_user',
'query': string,
}
# response:
{
'results': [('full_name', 'key', 'avatar_url'), ],
'status': 'OK',
'code': 200
}
"""
current.output = {
'results': [],
'status': 'OK',
'code': 201
}
qs = UserModel(current).objects.exclude(key=current.user_id).search_on(
*settings.MESSAGING_USER_SEARCH_FIELDS,
contains=current.input['query'])
# FIXME: somehow exclude(key=current.user_id) not working with search_on()
for user in qs:
if user.key != current.user_id:
current.output['results'].append((user.full_name, user.key, user.get_avatar_url()))
|
python
|
def search_user(current):
"""
Search users for adding to a public room
or creating one to one direct messaging
.. code-block:: python
# request:
{
'view':'_zops_search_user',
'query': string,
}
# response:
{
'results': [('full_name', 'key', 'avatar_url'), ],
'status': 'OK',
'code': 200
}
"""
current.output = {
'results': [],
'status': 'OK',
'code': 201
}
qs = UserModel(current).objects.exclude(key=current.user_id).search_on(
*settings.MESSAGING_USER_SEARCH_FIELDS,
contains=current.input['query'])
# FIXME: somehow exclude(key=current.user_id) not working with search_on()
for user in qs:
if user.key != current.user_id:
current.output['results'].append((user.full_name, user.key, user.get_avatar_url()))
|
[
"def",
"search_user",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'results'",
":",
"[",
"]",
",",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"201",
"}",
"qs",
"=",
"UserModel",
"(",
"current",
")",
".",
"objects",
".",
"exclude",
"(",
"key",
"=",
"current",
".",
"user_id",
")",
".",
"search_on",
"(",
"*",
"settings",
".",
"MESSAGING_USER_SEARCH_FIELDS",
",",
"contains",
"=",
"current",
".",
"input",
"[",
"'query'",
"]",
")",
"# FIXME: somehow exclude(key=current.user_id) not working with search_on()",
"for",
"user",
"in",
"qs",
":",
"if",
"user",
".",
"key",
"!=",
"current",
".",
"user_id",
":",
"current",
".",
"output",
"[",
"'results'",
"]",
".",
"append",
"(",
"(",
"user",
".",
"full_name",
",",
"user",
".",
"key",
",",
"user",
".",
"get_avatar_url",
"(",
")",
")",
")"
] |
Search users for adding to a public room
or creating one to one direct messaging
.. code-block:: python
# request:
{
'view':'_zops_search_user',
'query': string,
}
# response:
{
'results': [('full_name', 'key', 'avatar_url'), ],
'status': 'OK',
'code': 200
}
|
[
"Search",
"users",
"for",
"adding",
"to",
"a",
"public",
"room",
"or",
"creating",
"one",
"to",
"one",
"direct",
"messaging"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L531-L563
|
zetaops/zengine
|
zengine/messaging/views.py
|
search_unit
|
def search_unit(current):
"""
Search on units for subscribing it's users to a channel
.. code-block:: python
# request:
{
'view':'_zops_search_unit',
'query': string,
}
# response:
{
'results': [('name', 'key'), ],
'status': 'OK',
'code': 200
}
"""
current.output = {
'results': [],
'status': 'OK',
'code': 201
}
for user in UnitModel(current).objects.search_on(*settings.MESSAGING_UNIT_SEARCH_FIELDS,
contains=current.input['query']):
current.output['results'].append((user.name, user.key))
|
python
|
def search_unit(current):
"""
Search on units for subscribing it's users to a channel
.. code-block:: python
# request:
{
'view':'_zops_search_unit',
'query': string,
}
# response:
{
'results': [('name', 'key'), ],
'status': 'OK',
'code': 200
}
"""
current.output = {
'results': [],
'status': 'OK',
'code': 201
}
for user in UnitModel(current).objects.search_on(*settings.MESSAGING_UNIT_SEARCH_FIELDS,
contains=current.input['query']):
current.output['results'].append((user.name, user.key))
|
[
"def",
"search_unit",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'results'",
":",
"[",
"]",
",",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"201",
"}",
"for",
"user",
"in",
"UnitModel",
"(",
"current",
")",
".",
"objects",
".",
"search_on",
"(",
"*",
"settings",
".",
"MESSAGING_UNIT_SEARCH_FIELDS",
",",
"contains",
"=",
"current",
".",
"input",
"[",
"'query'",
"]",
")",
":",
"current",
".",
"output",
"[",
"'results'",
"]",
".",
"append",
"(",
"(",
"user",
".",
"name",
",",
"user",
".",
"key",
")",
")"
] |
Search on units for subscribing it's users to a channel
.. code-block:: python
# request:
{
'view':'_zops_search_unit',
'query': string,
}
# response:
{
'results': [('name', 'key'), ],
'status': 'OK',
'code': 200
}
|
[
"Search",
"on",
"units",
"for",
"subscribing",
"it",
"s",
"users",
"to",
"a",
"channel"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L566-L592
|
zetaops/zengine
|
zengine/messaging/views.py
|
create_direct_channel
|
def create_direct_channel(current):
"""
Create a One-To-One channel between current and selected user.
.. code-block:: python
# request:
{
'view':'_zops_create_direct_channel',
'user_key': key,
}
# response:
{
'description': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'last_messages': [MSG_DICT]
'status': 'Created',
'code': 201,
'channel_key': key, # of just created channel
'name': string, # name of subscribed channel
}
"""
channel, sub_name = Channel.get_or_create_direct_channel(current.user_id,
current.input['user_key'])
current.input['key'] = channel.key
show_channel(current)
current.output.update({
'status': 'Created',
'code': 201
})
|
python
|
def create_direct_channel(current):
"""
Create a One-To-One channel between current and selected user.
.. code-block:: python
# request:
{
'view':'_zops_create_direct_channel',
'user_key': key,
}
# response:
{
'description': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'last_messages': [MSG_DICT]
'status': 'Created',
'code': 201,
'channel_key': key, # of just created channel
'name': string, # name of subscribed channel
}
"""
channel, sub_name = Channel.get_or_create_direct_channel(current.user_id,
current.input['user_key'])
current.input['key'] = channel.key
show_channel(current)
current.output.update({
'status': 'Created',
'code': 201
})
|
[
"def",
"create_direct_channel",
"(",
"current",
")",
":",
"channel",
",",
"sub_name",
"=",
"Channel",
".",
"get_or_create_direct_channel",
"(",
"current",
".",
"user_id",
",",
"current",
".",
"input",
"[",
"'user_key'",
"]",
")",
"current",
".",
"input",
"[",
"'key'",
"]",
"=",
"channel",
".",
"key",
"show_channel",
"(",
"current",
")",
"current",
".",
"output",
".",
"update",
"(",
"{",
"'status'",
":",
"'Created'",
",",
"'code'",
":",
"201",
"}",
")"
] |
Create a One-To-One channel between current and selected user.
.. code-block:: python
# request:
{
'view':'_zops_create_direct_channel',
'user_key': key,
}
# response:
{
'description': string,
'no_of_members': int,
'member_list': [
{'name': string,
'is_online': bool,
'avatar_url': string,
}],
'last_messages': [MSG_DICT]
'status': 'Created',
'code': 201,
'channel_key': key, # of just created channel
'name': string, # name of subscribed channel
}
|
[
"Create",
"a",
"One",
"-",
"To",
"-",
"One",
"channel",
"between",
"current",
"and",
"selected",
"user",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L595-L631
|
zetaops/zengine
|
zengine/messaging/views.py
|
find_message
|
def find_message(current):
"""
Search in messages. If "channel_key" given, search will be limited to that channel,
otherwise search will be performed on all of user's subscribed channels.
.. code-block:: python
# request:
{
'view':'_zops_search_unit,
'channel_key': key,
'query': string,
'page': int,
}
# response:
{
'results': [MSG_DICT, ],
'pagination': {
'page': int, # current page
'total_pages': int,
'total_objects': int,
'per_page': int, # object per page
},
'status': 'OK',
'code': 200
}
"""
current.output = {
'results': [],
'status': 'OK',
'code': 201
}
query_set = Message(current).objects.search_on(['msg_title', 'body', 'url'],
contains=current.input['query'])
if current.input['channel_key']:
query_set = query_set.filter(channel_id=current.input['channel_key'])
else:
subscribed_channels = Subscriber.objects.filter(user_id=current.user_id).values_list(
"channel_id", flatten=True)
query_set = query_set.filter(channel_id__in=subscribed_channels)
query_set, pagination_data = _paginate(current_page=current.input['page'], query_set=query_set)
current.output['pagination'] = pagination_data
for msg in query_set:
current.output['results'].append(msg.serialize(current.user))
|
python
|
def find_message(current):
"""
Search in messages. If "channel_key" given, search will be limited to that channel,
otherwise search will be performed on all of user's subscribed channels.
.. code-block:: python
# request:
{
'view':'_zops_search_unit,
'channel_key': key,
'query': string,
'page': int,
}
# response:
{
'results': [MSG_DICT, ],
'pagination': {
'page': int, # current page
'total_pages': int,
'total_objects': int,
'per_page': int, # object per page
},
'status': 'OK',
'code': 200
}
"""
current.output = {
'results': [],
'status': 'OK',
'code': 201
}
query_set = Message(current).objects.search_on(['msg_title', 'body', 'url'],
contains=current.input['query'])
if current.input['channel_key']:
query_set = query_set.filter(channel_id=current.input['channel_key'])
else:
subscribed_channels = Subscriber.objects.filter(user_id=current.user_id).values_list(
"channel_id", flatten=True)
query_set = query_set.filter(channel_id__in=subscribed_channels)
query_set, pagination_data = _paginate(current_page=current.input['page'], query_set=query_set)
current.output['pagination'] = pagination_data
for msg in query_set:
current.output['results'].append(msg.serialize(current.user))
|
[
"def",
"find_message",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'results'",
":",
"[",
"]",
",",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"201",
"}",
"query_set",
"=",
"Message",
"(",
"current",
")",
".",
"objects",
".",
"search_on",
"(",
"[",
"'msg_title'",
",",
"'body'",
",",
"'url'",
"]",
",",
"contains",
"=",
"current",
".",
"input",
"[",
"'query'",
"]",
")",
"if",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
":",
"query_set",
"=",
"query_set",
".",
"filter",
"(",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
")",
"else",
":",
"subscribed_channels",
"=",
"Subscriber",
".",
"objects",
".",
"filter",
"(",
"user_id",
"=",
"current",
".",
"user_id",
")",
".",
"values_list",
"(",
"\"channel_id\"",
",",
"flatten",
"=",
"True",
")",
"query_set",
"=",
"query_set",
".",
"filter",
"(",
"channel_id__in",
"=",
"subscribed_channels",
")",
"query_set",
",",
"pagination_data",
"=",
"_paginate",
"(",
"current_page",
"=",
"current",
".",
"input",
"[",
"'page'",
"]",
",",
"query_set",
"=",
"query_set",
")",
"current",
".",
"output",
"[",
"'pagination'",
"]",
"=",
"pagination_data",
"for",
"msg",
"in",
"query_set",
":",
"current",
".",
"output",
"[",
"'results'",
"]",
".",
"append",
"(",
"msg",
".",
"serialize",
"(",
"current",
".",
"user",
")",
")"
] |
Search in messages. If "channel_key" given, search will be limited to that channel,
otherwise search will be performed on all of user's subscribed channels.
.. code-block:: python
# request:
{
'view':'_zops_search_unit,
'channel_key': key,
'query': string,
'page': int,
}
# response:
{
'results': [MSG_DICT, ],
'pagination': {
'page': int, # current page
'total_pages': int,
'total_objects': int,
'per_page': int, # object per page
},
'status': 'OK',
'code': 200
}
|
[
"Search",
"in",
"messages",
".",
"If",
"channel_key",
"given",
"search",
"will",
"be",
"limited",
"to",
"that",
"channel",
"otherwise",
"search",
"will",
"be",
"performed",
"on",
"all",
"of",
"user",
"s",
"subscribed",
"channels",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L634-L679
|
zetaops/zengine
|
zengine/messaging/views.py
|
delete_channel
|
def delete_channel(current):
"""
Delete a channel
.. code-block:: python
# request:
{
'view':'_zops_delete_channel,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
ch_key = current.input['channel_key']
ch = Channel(current).objects.get(owner_id=current.user_id, key=ch_key)
ch.delete()
Subscriber.objects.filter(channel_id=ch_key).delete()
Message.objects.filter(channel_id=ch_key).delete()
current.output = {'status': 'Deleted', 'code': 200}
|
python
|
def delete_channel(current):
"""
Delete a channel
.. code-block:: python
# request:
{
'view':'_zops_delete_channel,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
ch_key = current.input['channel_key']
ch = Channel(current).objects.get(owner_id=current.user_id, key=ch_key)
ch.delete()
Subscriber.objects.filter(channel_id=ch_key).delete()
Message.objects.filter(channel_id=ch_key).delete()
current.output = {'status': 'Deleted', 'code': 200}
|
[
"def",
"delete_channel",
"(",
"current",
")",
":",
"ch_key",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
"ch",
"=",
"Channel",
"(",
"current",
")",
".",
"objects",
".",
"get",
"(",
"owner_id",
"=",
"current",
".",
"user_id",
",",
"key",
"=",
"ch_key",
")",
"ch",
".",
"delete",
"(",
")",
"Subscriber",
".",
"objects",
".",
"filter",
"(",
"channel_id",
"=",
"ch_key",
")",
".",
"delete",
"(",
")",
"Message",
".",
"objects",
".",
"filter",
"(",
"channel_id",
"=",
"ch_key",
")",
".",
"delete",
"(",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'Deleted'",
",",
"'code'",
":",
"200",
"}"
] |
Delete a channel
.. code-block:: python
# request:
{
'view':'_zops_delete_channel,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
|
[
"Delete",
"a",
"channel"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L682-L706
|
zetaops/zengine
|
zengine/messaging/views.py
|
edit_channel
|
def edit_channel(current):
"""
Update channel name or description
.. code-block:: python
# request:
{
'view':'_zops_edit_channel,
'channel_key': key,
'name': string,
'description': string,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
ch = Channel(current).objects.get(owner_id=current.user_id,
key=current.input['channel_key'])
ch.name = current.input['name']
ch.description = current.input['description']
ch.save()
for sbs in ch.subscriber_set.objects.all():
sbs.name = ch.name
sbs.save()
current.output = {'status': 'OK', 'code': 200}
|
python
|
def edit_channel(current):
"""
Update channel name or description
.. code-block:: python
# request:
{
'view':'_zops_edit_channel,
'channel_key': key,
'name': string,
'description': string,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
ch = Channel(current).objects.get(owner_id=current.user_id,
key=current.input['channel_key'])
ch.name = current.input['name']
ch.description = current.input['description']
ch.save()
for sbs in ch.subscriber_set.objects.all():
sbs.name = ch.name
sbs.save()
current.output = {'status': 'OK', 'code': 200}
|
[
"def",
"edit_channel",
"(",
"current",
")",
":",
"ch",
"=",
"Channel",
"(",
"current",
")",
".",
"objects",
".",
"get",
"(",
"owner_id",
"=",
"current",
".",
"user_id",
",",
"key",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
")",
"ch",
".",
"name",
"=",
"current",
".",
"input",
"[",
"'name'",
"]",
"ch",
".",
"description",
"=",
"current",
".",
"input",
"[",
"'description'",
"]",
"ch",
".",
"save",
"(",
")",
"for",
"sbs",
"in",
"ch",
".",
"subscriber_set",
".",
"objects",
".",
"all",
"(",
")",
":",
"sbs",
".",
"name",
"=",
"ch",
".",
"name",
"sbs",
".",
"save",
"(",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}"
] |
Update channel name or description
.. code-block:: python
# request:
{
'view':'_zops_edit_channel,
'channel_key': key,
'name': string,
'description': string,
}
# response:
{
'status': 'OK',
'code': 200
}
|
[
"Update",
"channel",
"name",
"or",
"description"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L709-L737
|
zetaops/zengine
|
zengine/messaging/views.py
|
pin_channel
|
def pin_channel(current):
"""
Pin a channel to top of channel list
.. code-block:: python
# request:
{
'view':'_zops_pin_channel,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
try:
Subscriber(current).objects.filter(user_id=current.user_id,
channel_id=current.input['channel_key']).update(
pinned=True)
current.output = {'status': 'OK', 'code': 200}
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
python
|
def pin_channel(current):
"""
Pin a channel to top of channel list
.. code-block:: python
# request:
{
'view':'_zops_pin_channel,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
try:
Subscriber(current).objects.filter(user_id=current.user_id,
channel_id=current.input['channel_key']).update(
pinned=True)
current.output = {'status': 'OK', 'code': 200}
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
[
"def",
"pin_channel",
"(",
"current",
")",
":",
"try",
":",
"Subscriber",
"(",
"current",
")",
".",
"objects",
".",
"filter",
"(",
"user_id",
"=",
"current",
".",
"user_id",
",",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
")",
".",
"update",
"(",
"pinned",
"=",
"True",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}",
"except",
"ObjectDoesNotExist",
":",
"raise",
"HTTPError",
"(",
"404",
",",
"\"\"",
")"
] |
Pin a channel to top of channel list
.. code-block:: python
# request:
{
'view':'_zops_pin_channel,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
|
[
"Pin",
"a",
"channel",
"to",
"top",
"of",
"channel",
"list"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L740-L764
|
zetaops/zengine
|
zengine/messaging/views.py
|
delete_message
|
def delete_message(current):
"""
Delete a message
.. code-block:: python
# request:
{
'view':'_zops_delete_message,
'message_key': key,
}
# response:
{
'key': key,
'status': 'OK',
'code': 200
}
"""
try:
Message(current).objects.get(sender_id=current.user_id,
key=current.input['key']).delete()
current.output = {'status': 'Deleted', 'code': 200, 'key': current.input['key']}
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
python
|
def delete_message(current):
"""
Delete a message
.. code-block:: python
# request:
{
'view':'_zops_delete_message,
'message_key': key,
}
# response:
{
'key': key,
'status': 'OK',
'code': 200
}
"""
try:
Message(current).objects.get(sender_id=current.user_id,
key=current.input['key']).delete()
current.output = {'status': 'Deleted', 'code': 200, 'key': current.input['key']}
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
[
"def",
"delete_message",
"(",
"current",
")",
":",
"try",
":",
"Message",
"(",
"current",
")",
".",
"objects",
".",
"get",
"(",
"sender_id",
"=",
"current",
".",
"user_id",
",",
"key",
"=",
"current",
".",
"input",
"[",
"'key'",
"]",
")",
".",
"delete",
"(",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'Deleted'",
",",
"'code'",
":",
"200",
",",
"'key'",
":",
"current",
".",
"input",
"[",
"'key'",
"]",
"}",
"except",
"ObjectDoesNotExist",
":",
"raise",
"HTTPError",
"(",
"404",
",",
"\"\"",
")"
] |
Delete a message
.. code-block:: python
# request:
{
'view':'_zops_delete_message,
'message_key': key,
}
# response:
{
'key': key,
'status': 'OK',
'code': 200
}
|
[
"Delete",
"a",
"message"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L767-L791
|
zetaops/zengine
|
zengine/messaging/views.py
|
edit_message
|
def edit_message(current):
"""
Edit a message a user own.
.. code-block:: python
# request:
{
'view':'_zops_edit_message',
'message': {
'body': string, # message text
'key': key
}
}
# response:
{
'status': string, # 'OK' for success
'code': int, # 200 for success
}
"""
current.output = {'status': 'OK', 'code': 200}
in_msg = current.input['message']
try:
msg = Message(current).objects.get(sender_id=current.user_id, key=in_msg['key'])
msg.body = in_msg['body']
msg.save()
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
python
|
def edit_message(current):
"""
Edit a message a user own.
.. code-block:: python
# request:
{
'view':'_zops_edit_message',
'message': {
'body': string, # message text
'key': key
}
}
# response:
{
'status': string, # 'OK' for success
'code': int, # 200 for success
}
"""
current.output = {'status': 'OK', 'code': 200}
in_msg = current.input['message']
try:
msg = Message(current).objects.get(sender_id=current.user_id, key=in_msg['key'])
msg.body = in_msg['body']
msg.save()
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
[
"def",
"edit_message",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}",
"in_msg",
"=",
"current",
".",
"input",
"[",
"'message'",
"]",
"try",
":",
"msg",
"=",
"Message",
"(",
"current",
")",
".",
"objects",
".",
"get",
"(",
"sender_id",
"=",
"current",
".",
"user_id",
",",
"key",
"=",
"in_msg",
"[",
"'key'",
"]",
")",
"msg",
".",
"body",
"=",
"in_msg",
"[",
"'body'",
"]",
"msg",
".",
"save",
"(",
")",
"except",
"ObjectDoesNotExist",
":",
"raise",
"HTTPError",
"(",
"404",
",",
"\"\"",
")"
] |
Edit a message a user own.
.. code-block:: python
# request:
{
'view':'_zops_edit_message',
'message': {
'body': string, # message text
'key': key
}
}
# response:
{
'status': string, # 'OK' for success
'code': int, # 200 for success
}
|
[
"Edit",
"a",
"message",
"a",
"user",
"own",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L794-L822
|
zetaops/zengine
|
zengine/messaging/views.py
|
flag_message
|
def flag_message(current):
"""
Flag inappropriate messages
.. code-block:: python
# request:
{
'view':'_zops_flag_message',
'message_key': key,
}
# response:
{
'
'status': 'Created',
'code': 201,
}
"""
current.output = {'status': 'Created', 'code': 201}
FlaggedMessage.objects.get_or_create(user_id=current.user_id,
message_id=current.input['key'])
|
python
|
def flag_message(current):
"""
Flag inappropriate messages
.. code-block:: python
# request:
{
'view':'_zops_flag_message',
'message_key': key,
}
# response:
{
'
'status': 'Created',
'code': 201,
}
"""
current.output = {'status': 'Created', 'code': 201}
FlaggedMessage.objects.get_or_create(user_id=current.user_id,
message_id=current.input['key'])
|
[
"def",
"flag_message",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'Created'",
",",
"'code'",
":",
"201",
"}",
"FlaggedMessage",
".",
"objects",
".",
"get_or_create",
"(",
"user_id",
"=",
"current",
".",
"user_id",
",",
"message_id",
"=",
"current",
".",
"input",
"[",
"'key'",
"]",
")"
] |
Flag inappropriate messages
.. code-block:: python
# request:
{
'view':'_zops_flag_message',
'message_key': key,
}
# response:
{
'
'status': 'Created',
'code': 201,
}
|
[
"Flag",
"inappropriate",
"messages"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L825-L846
|
zetaops/zengine
|
zengine/messaging/views.py
|
unflag_message
|
def unflag_message(current):
"""
remove flag of a message
.. code-block:: python
# request:
{
'view':'_zops_flag_message',
'key': key,
}
# response:
{
'
'status': 'OK',
'code': 200,
}
"""
current.output = {'status': 'OK', 'code': 200}
FlaggedMessage(current).objects.filter(user_id=current.user_id,
message_id=current.input['key']).delete()
|
python
|
def unflag_message(current):
"""
remove flag of a message
.. code-block:: python
# request:
{
'view':'_zops_flag_message',
'key': key,
}
# response:
{
'
'status': 'OK',
'code': 200,
}
"""
current.output = {'status': 'OK', 'code': 200}
FlaggedMessage(current).objects.filter(user_id=current.user_id,
message_id=current.input['key']).delete()
|
[
"def",
"unflag_message",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}",
"FlaggedMessage",
"(",
"current",
")",
".",
"objects",
".",
"filter",
"(",
"user_id",
"=",
"current",
".",
"user_id",
",",
"message_id",
"=",
"current",
".",
"input",
"[",
"'key'",
"]",
")",
".",
"delete",
"(",
")"
] |
remove flag of a message
.. code-block:: python
# request:
{
'view':'_zops_flag_message',
'key': key,
}
# response:
{
'
'status': 'OK',
'code': 200,
}
|
[
"remove",
"flag",
"of",
"a",
"message"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L849-L871
|
zetaops/zengine
|
zengine/messaging/views.py
|
get_message_actions
|
def get_message_actions(current):
"""
Returns applicable actions for current user for given message key
.. code-block:: python
# request:
{
'view':'_zops_get_message_actions',
'key': key,
}
# response:
{
'actions':[('name_string', 'cmd_string'),]
'status': string, # 'OK' for success
'code': int, # 200 for success
}
"""
current.output = {'status': 'OK',
'code': 200,
'actions': Message.objects.get(
current.input['key']).get_actions_for(current.user)}
|
python
|
def get_message_actions(current):
"""
Returns applicable actions for current user for given message key
.. code-block:: python
# request:
{
'view':'_zops_get_message_actions',
'key': key,
}
# response:
{
'actions':[('name_string', 'cmd_string'),]
'status': string, # 'OK' for success
'code': int, # 200 for success
}
"""
current.output = {'status': 'OK',
'code': 200,
'actions': Message.objects.get(
current.input['key']).get_actions_for(current.user)}
|
[
"def",
"get_message_actions",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
",",
"'actions'",
":",
"Message",
".",
"objects",
".",
"get",
"(",
"current",
".",
"input",
"[",
"'key'",
"]",
")",
".",
"get_actions_for",
"(",
"current",
".",
"user",
")",
"}"
] |
Returns applicable actions for current user for given message key
.. code-block:: python
# request:
{
'view':'_zops_get_message_actions',
'key': key,
}
# response:
{
'actions':[('name_string', 'cmd_string'),]
'status': string, # 'OK' for success
'code': int, # 200 for success
}
|
[
"Returns",
"applicable",
"actions",
"for",
"current",
"user",
"for",
"given",
"message",
"key"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L874-L896
|
zetaops/zengine
|
zengine/messaging/views.py
|
add_to_favorites
|
def add_to_favorites(current):
"""
Favorite a message
.. code-block:: python
# request:
{
'view':'_zops_add_to_favorites,
'key': key,
}
# response:
{
'status': 'Created',
'code': 201
'favorite_key': key
}
"""
msg = Message.objects.get(current.input['key'])
current.output = {'status': 'Created', 'code': 201}
fav, new = Favorite.objects.get_or_create(user_id=current.user_id, message=msg)
current.output['favorite_key'] = fav.key
|
python
|
def add_to_favorites(current):
"""
Favorite a message
.. code-block:: python
# request:
{
'view':'_zops_add_to_favorites,
'key': key,
}
# response:
{
'status': 'Created',
'code': 201
'favorite_key': key
}
"""
msg = Message.objects.get(current.input['key'])
current.output = {'status': 'Created', 'code': 201}
fav, new = Favorite.objects.get_or_create(user_id=current.user_id, message=msg)
current.output['favorite_key'] = fav.key
|
[
"def",
"add_to_favorites",
"(",
"current",
")",
":",
"msg",
"=",
"Message",
".",
"objects",
".",
"get",
"(",
"current",
".",
"input",
"[",
"'key'",
"]",
")",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'Created'",
",",
"'code'",
":",
"201",
"}",
"fav",
",",
"new",
"=",
"Favorite",
".",
"objects",
".",
"get_or_create",
"(",
"user_id",
"=",
"current",
".",
"user_id",
",",
"message",
"=",
"msg",
")",
"current",
".",
"output",
"[",
"'favorite_key'",
"]",
"=",
"fav",
".",
"key"
] |
Favorite a message
.. code-block:: python
# request:
{
'view':'_zops_add_to_favorites,
'key': key,
}
# response:
{
'status': 'Created',
'code': 201
'favorite_key': key
}
|
[
"Favorite",
"a",
"message"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L899-L922
|
zetaops/zengine
|
zengine/messaging/views.py
|
remove_from_favorites
|
def remove_from_favorites(current):
"""
Remove a message from favorites
.. code-block:: python
# request:
{
'view':'_zops_remove_from_favorites,
'key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
try:
current.output = {'status': 'OK', 'code': 200}
Favorite(current).objects.get(user_id=current.user_id,
key=current.input['key']).delete()
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
python
|
def remove_from_favorites(current):
"""
Remove a message from favorites
.. code-block:: python
# request:
{
'view':'_zops_remove_from_favorites,
'key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
"""
try:
current.output = {'status': 'OK', 'code': 200}
Favorite(current).objects.get(user_id=current.user_id,
key=current.input['key']).delete()
except ObjectDoesNotExist:
raise HTTPError(404, "")
|
[
"def",
"remove_from_favorites",
"(",
"current",
")",
":",
"try",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
"}",
"Favorite",
"(",
"current",
")",
".",
"objects",
".",
"get",
"(",
"user_id",
"=",
"current",
".",
"user_id",
",",
"key",
"=",
"current",
".",
"input",
"[",
"'key'",
"]",
")",
".",
"delete",
"(",
")",
"except",
"ObjectDoesNotExist",
":",
"raise",
"HTTPError",
"(",
"404",
",",
"\"\"",
")"
] |
Remove a message from favorites
.. code-block:: python
# request:
{
'view':'_zops_remove_from_favorites,
'key': key,
}
# response:
{
'status': 'OK',
'code': 200
}
|
[
"Remove",
"a",
"message",
"from",
"favorites"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L925-L949
|
zetaops/zengine
|
zengine/messaging/views.py
|
list_favorites
|
def list_favorites(current):
"""
List user's favorites. If "channel_key" given, will return favorites belong to that channel.
.. code-block:: python
# request:
{
'view':'_zops_list_favorites,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
'favorites':[{'key': key,
'channel_key': key,
'message_key': key,
'message_summary': string, # max 60 char
'channel_name': string,
},]
}
"""
current.output = {'status': 'OK', 'code': 200, 'favorites': []}
query_set = Favorite(current).objects.filter(user_id=current.user_id)
if current.input['channel_key']:
query_set = query_set.filter(channel_id=current.input['channel_key'])
current.output['favorites'] = [{
'key': fav.key,
'channel_key': fav.channel.key,
'message_key': fav.message.key,
'message_summary': fav.summary,
'channel_name': fav.channel_name
} for fav in query_set]
|
python
|
def list_favorites(current):
"""
List user's favorites. If "channel_key" given, will return favorites belong to that channel.
.. code-block:: python
# request:
{
'view':'_zops_list_favorites,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
'favorites':[{'key': key,
'channel_key': key,
'message_key': key,
'message_summary': string, # max 60 char
'channel_name': string,
},]
}
"""
current.output = {'status': 'OK', 'code': 200, 'favorites': []}
query_set = Favorite(current).objects.filter(user_id=current.user_id)
if current.input['channel_key']:
query_set = query_set.filter(channel_id=current.input['channel_key'])
current.output['favorites'] = [{
'key': fav.key,
'channel_key': fav.channel.key,
'message_key': fav.message.key,
'message_summary': fav.summary,
'channel_name': fav.channel_name
} for fav in query_set]
|
[
"def",
"list_favorites",
"(",
"current",
")",
":",
"current",
".",
"output",
"=",
"{",
"'status'",
":",
"'OK'",
",",
"'code'",
":",
"200",
",",
"'favorites'",
":",
"[",
"]",
"}",
"query_set",
"=",
"Favorite",
"(",
"current",
")",
".",
"objects",
".",
"filter",
"(",
"user_id",
"=",
"current",
".",
"user_id",
")",
"if",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
":",
"query_set",
"=",
"query_set",
".",
"filter",
"(",
"channel_id",
"=",
"current",
".",
"input",
"[",
"'channel_key'",
"]",
")",
"current",
".",
"output",
"[",
"'favorites'",
"]",
"=",
"[",
"{",
"'key'",
":",
"fav",
".",
"key",
",",
"'channel_key'",
":",
"fav",
".",
"channel",
".",
"key",
",",
"'message_key'",
":",
"fav",
".",
"message",
".",
"key",
",",
"'message_summary'",
":",
"fav",
".",
"summary",
",",
"'channel_name'",
":",
"fav",
".",
"channel_name",
"}",
"for",
"fav",
"in",
"query_set",
"]"
] |
List user's favorites. If "channel_key" given, will return favorites belong to that channel.
.. code-block:: python
# request:
{
'view':'_zops_list_favorites,
'channel_key': key,
}
# response:
{
'status': 'OK',
'code': 200
'favorites':[{'key': key,
'channel_key': key,
'message_key': key,
'message_summary': string, # max 60 char
'channel_name': string,
},]
}
|
[
"List",
"user",
"s",
"favorites",
".",
"If",
"channel_key",
"given",
"will",
"return",
"favorites",
"belong",
"to",
"that",
"channel",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/views.py#L952-L987
|
zetaops/zengine
|
zengine/messaging/model.py
|
Channel.get_or_create_direct_channel
|
def get_or_create_direct_channel(cls, initiator_key, receiver_key):
"""
Creates a direct messaging channel between two user
Args:
initiator: User, who want's to make first contact
receiver: User, other party
Returns:
(Channel, receiver_name)
"""
existing = cls.objects.OR().filter(
code_name='%s_%s' % (initiator_key, receiver_key)).filter(
code_name='%s_%s' % (receiver_key, initiator_key))
receiver_name = UserModel.objects.get(receiver_key).full_name
if existing:
channel = existing[0]
else:
channel_name = '%s_%s' % (initiator_key, receiver_key)
channel = cls(is_direct=True, code_name=channel_name, typ=10).blocking_save()
with BlockSave(Subscriber):
Subscriber.objects.get_or_create(channel=channel,
user_id=initiator_key,
name=receiver_name)
Subscriber.objects.get_or_create(channel=channel,
user_id=receiver_key,
name=UserModel.objects.get(initiator_key).full_name)
return channel, receiver_name
|
python
|
def get_or_create_direct_channel(cls, initiator_key, receiver_key):
"""
Creates a direct messaging channel between two user
Args:
initiator: User, who want's to make first contact
receiver: User, other party
Returns:
(Channel, receiver_name)
"""
existing = cls.objects.OR().filter(
code_name='%s_%s' % (initiator_key, receiver_key)).filter(
code_name='%s_%s' % (receiver_key, initiator_key))
receiver_name = UserModel.objects.get(receiver_key).full_name
if existing:
channel = existing[0]
else:
channel_name = '%s_%s' % (initiator_key, receiver_key)
channel = cls(is_direct=True, code_name=channel_name, typ=10).blocking_save()
with BlockSave(Subscriber):
Subscriber.objects.get_or_create(channel=channel,
user_id=initiator_key,
name=receiver_name)
Subscriber.objects.get_or_create(channel=channel,
user_id=receiver_key,
name=UserModel.objects.get(initiator_key).full_name)
return channel, receiver_name
|
[
"def",
"get_or_create_direct_channel",
"(",
"cls",
",",
"initiator_key",
",",
"receiver_key",
")",
":",
"existing",
"=",
"cls",
".",
"objects",
".",
"OR",
"(",
")",
".",
"filter",
"(",
"code_name",
"=",
"'%s_%s'",
"%",
"(",
"initiator_key",
",",
"receiver_key",
")",
")",
".",
"filter",
"(",
"code_name",
"=",
"'%s_%s'",
"%",
"(",
"receiver_key",
",",
"initiator_key",
")",
")",
"receiver_name",
"=",
"UserModel",
".",
"objects",
".",
"get",
"(",
"receiver_key",
")",
".",
"full_name",
"if",
"existing",
":",
"channel",
"=",
"existing",
"[",
"0",
"]",
"else",
":",
"channel_name",
"=",
"'%s_%s'",
"%",
"(",
"initiator_key",
",",
"receiver_key",
")",
"channel",
"=",
"cls",
"(",
"is_direct",
"=",
"True",
",",
"code_name",
"=",
"channel_name",
",",
"typ",
"=",
"10",
")",
".",
"blocking_save",
"(",
")",
"with",
"BlockSave",
"(",
"Subscriber",
")",
":",
"Subscriber",
".",
"objects",
".",
"get_or_create",
"(",
"channel",
"=",
"channel",
",",
"user_id",
"=",
"initiator_key",
",",
"name",
"=",
"receiver_name",
")",
"Subscriber",
".",
"objects",
".",
"get_or_create",
"(",
"channel",
"=",
"channel",
",",
"user_id",
"=",
"receiver_key",
",",
"name",
"=",
"UserModel",
".",
"objects",
".",
"get",
"(",
"initiator_key",
")",
".",
"full_name",
")",
"return",
"channel",
",",
"receiver_name"
] |
Creates a direct messaging channel between two user
Args:
initiator: User, who want's to make first contact
receiver: User, other party
Returns:
(Channel, receiver_name)
|
[
"Creates",
"a",
"direct",
"messaging",
"channel",
"between",
"two",
"user"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L75-L102
|
zetaops/zengine
|
zengine/messaging/model.py
|
Channel.create_exchange
|
def create_exchange(self):
"""
Creates MQ exchange for this channel
Needs to be defined only once.
"""
mq_channel = self._connect_mq()
mq_channel.exchange_declare(exchange=self.code_name,
exchange_type='fanout',
durable=True)
|
python
|
def create_exchange(self):
"""
Creates MQ exchange for this channel
Needs to be defined only once.
"""
mq_channel = self._connect_mq()
mq_channel.exchange_declare(exchange=self.code_name,
exchange_type='fanout',
durable=True)
|
[
"def",
"create_exchange",
"(",
"self",
")",
":",
"mq_channel",
"=",
"self",
".",
"_connect_mq",
"(",
")",
"mq_channel",
".",
"exchange_declare",
"(",
"exchange",
"=",
"self",
".",
"code_name",
",",
"exchange_type",
"=",
"'fanout'",
",",
"durable",
"=",
"True",
")"
] |
Creates MQ exchange for this channel
Needs to be defined only once.
|
[
"Creates",
"MQ",
"exchange",
"for",
"this",
"channel",
"Needs",
"to",
"be",
"defined",
"only",
"once",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L135-L143
|
zetaops/zengine
|
zengine/messaging/model.py
|
Channel.delete_exchange
|
def delete_exchange(self):
"""
Deletes MQ exchange for this channel
Needs to be defined only once.
"""
mq_channel = self._connect_mq()
mq_channel.exchange_delete(exchange=self.code_name)
|
python
|
def delete_exchange(self):
"""
Deletes MQ exchange for this channel
Needs to be defined only once.
"""
mq_channel = self._connect_mq()
mq_channel.exchange_delete(exchange=self.code_name)
|
[
"def",
"delete_exchange",
"(",
"self",
")",
":",
"mq_channel",
"=",
"self",
".",
"_connect_mq",
"(",
")",
"mq_channel",
".",
"exchange_delete",
"(",
"exchange",
"=",
"self",
".",
"code_name",
")"
] |
Deletes MQ exchange for this channel
Needs to be defined only once.
|
[
"Deletes",
"MQ",
"exchange",
"for",
"this",
"channel",
"Needs",
"to",
"be",
"defined",
"only",
"once",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L145-L151
|
zetaops/zengine
|
zengine/messaging/model.py
|
Subscriber.get_channel_listing
|
def get_channel_listing(self):
"""
serialized form for channel listing
"""
return {'name': self.name,
'key': self.channel.key,
'type': self.channel.typ,
'read_only': self.read_only,
'is_online': self.is_online(),
'actions': self.get_actions(),
'unread': self.unread_count()}
|
python
|
def get_channel_listing(self):
"""
serialized form for channel listing
"""
return {'name': self.name,
'key': self.channel.key,
'type': self.channel.typ,
'read_only': self.read_only,
'is_online': self.is_online(),
'actions': self.get_actions(),
'unread': self.unread_count()}
|
[
"def",
"get_channel_listing",
"(",
"self",
")",
":",
"return",
"{",
"'name'",
":",
"self",
".",
"name",
",",
"'key'",
":",
"self",
".",
"channel",
".",
"key",
",",
"'type'",
":",
"self",
".",
"channel",
".",
"typ",
",",
"'read_only'",
":",
"self",
".",
"read_only",
",",
"'is_online'",
":",
"self",
".",
"is_online",
"(",
")",
",",
"'actions'",
":",
"self",
".",
"get_actions",
"(",
")",
",",
"'unread'",
":",
"self",
".",
"unread_count",
"(",
")",
"}"
] |
serialized form for channel listing
|
[
"serialized",
"form",
"for",
"channel",
"listing"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L214-L225
|
zetaops/zengine
|
zengine/messaging/model.py
|
Subscriber.create_exchange
|
def create_exchange(self):
"""
Creates user's private exchange
Actually user's private channel needed to be defined only once,
and this should be happened when user first created.
But since this has a little performance cost,
to be safe we always call it before binding to the channel we currently subscribe
"""
channel = self._connect_mq()
channel.exchange_declare(exchange=self.user.prv_exchange,
exchange_type='fanout',
durable=True)
|
python
|
def create_exchange(self):
"""
Creates user's private exchange
Actually user's private channel needed to be defined only once,
and this should be happened when user first created.
But since this has a little performance cost,
to be safe we always call it before binding to the channel we currently subscribe
"""
channel = self._connect_mq()
channel.exchange_declare(exchange=self.user.prv_exchange,
exchange_type='fanout',
durable=True)
|
[
"def",
"create_exchange",
"(",
"self",
")",
":",
"channel",
"=",
"self",
".",
"_connect_mq",
"(",
")",
"channel",
".",
"exchange_declare",
"(",
"exchange",
"=",
"self",
".",
"user",
".",
"prv_exchange",
",",
"exchange_type",
"=",
"'fanout'",
",",
"durable",
"=",
"True",
")"
] |
Creates user's private exchange
Actually user's private channel needed to be defined only once,
and this should be happened when user first created.
But since this has a little performance cost,
to be safe we always call it before binding to the channel we currently subscribe
|
[
"Creates",
"user",
"s",
"private",
"exchange"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L264-L276
|
zetaops/zengine
|
zengine/messaging/model.py
|
Subscriber.bind_to_channel
|
def bind_to_channel(self):
"""
Binds (subscribes) users private exchange to channel exchange
Automatically called at creation of subscription record.
"""
if self.channel.code_name != self.user.prv_exchange:
channel = self._connect_mq()
channel.exchange_bind(source=self.channel.code_name, destination=self.user.prv_exchange)
|
python
|
def bind_to_channel(self):
"""
Binds (subscribes) users private exchange to channel exchange
Automatically called at creation of subscription record.
"""
if self.channel.code_name != self.user.prv_exchange:
channel = self._connect_mq()
channel.exchange_bind(source=self.channel.code_name, destination=self.user.prv_exchange)
|
[
"def",
"bind_to_channel",
"(",
"self",
")",
":",
"if",
"self",
".",
"channel",
".",
"code_name",
"!=",
"self",
".",
"user",
".",
"prv_exchange",
":",
"channel",
"=",
"self",
".",
"_connect_mq",
"(",
")",
"channel",
".",
"exchange_bind",
"(",
"source",
"=",
"self",
".",
"channel",
".",
"code_name",
",",
"destination",
"=",
"self",
".",
"user",
".",
"prv_exchange",
")"
] |
Binds (subscribes) users private exchange to channel exchange
Automatically called at creation of subscription record.
|
[
"Binds",
"(",
"subscribes",
")",
"users",
"private",
"exchange",
"to",
"channel",
"exchange",
"Automatically",
"called",
"at",
"creation",
"of",
"subscription",
"record",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L282-L289
|
zetaops/zengine
|
zengine/messaging/model.py
|
Message.serialize
|
def serialize(self, user=None):
"""
Serializes message for given user.
Note:
Should be called before first save(). Otherwise "is_update" will get wrong value.
Args:
user: User object
Returns:
Dict. JSON serialization ready dictionary object
"""
return {
'content': self.body,
'type': self.typ,
'updated_at': self.updated_at,
'timestamp': self.updated_at,
'is_update': not hasattr(self, 'unsaved'),
'attachments': [attachment.serialize() for attachment in self.attachment_set],
'title': self.msg_title,
'url': self.url,
'sender_name': self.sender.full_name,
'sender_key': self.sender.key,
'channel_key': self.channel.key,
'cmd': 'message',
'avatar_url': self.sender.avatar,
'key': self.key,
}
|
python
|
def serialize(self, user=None):
"""
Serializes message for given user.
Note:
Should be called before first save(). Otherwise "is_update" will get wrong value.
Args:
user: User object
Returns:
Dict. JSON serialization ready dictionary object
"""
return {
'content': self.body,
'type': self.typ,
'updated_at': self.updated_at,
'timestamp': self.updated_at,
'is_update': not hasattr(self, 'unsaved'),
'attachments': [attachment.serialize() for attachment in self.attachment_set],
'title': self.msg_title,
'url': self.url,
'sender_name': self.sender.full_name,
'sender_key': self.sender.key,
'channel_key': self.channel.key,
'cmd': 'message',
'avatar_url': self.sender.avatar,
'key': self.key,
}
|
[
"def",
"serialize",
"(",
"self",
",",
"user",
"=",
"None",
")",
":",
"return",
"{",
"'content'",
":",
"self",
".",
"body",
",",
"'type'",
":",
"self",
".",
"typ",
",",
"'updated_at'",
":",
"self",
".",
"updated_at",
",",
"'timestamp'",
":",
"self",
".",
"updated_at",
",",
"'is_update'",
":",
"not",
"hasattr",
"(",
"self",
",",
"'unsaved'",
")",
",",
"'attachments'",
":",
"[",
"attachment",
".",
"serialize",
"(",
")",
"for",
"attachment",
"in",
"self",
".",
"attachment_set",
"]",
",",
"'title'",
":",
"self",
".",
"msg_title",
",",
"'url'",
":",
"self",
".",
"url",
",",
"'sender_name'",
":",
"self",
".",
"sender",
".",
"full_name",
",",
"'sender_key'",
":",
"self",
".",
"sender",
".",
"key",
",",
"'channel_key'",
":",
"self",
".",
"channel",
".",
"key",
",",
"'cmd'",
":",
"'message'",
",",
"'avatar_url'",
":",
"self",
".",
"sender",
".",
"avatar",
",",
"'key'",
":",
"self",
".",
"key",
",",
"}"
] |
Serializes message for given user.
Note:
Should be called before first save(). Otherwise "is_update" will get wrong value.
Args:
user: User object
Returns:
Dict. JSON serialization ready dictionary object
|
[
"Serializes",
"message",
"for",
"given",
"user",
"."
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L370-L398
|
zetaops/zengine
|
zengine/messaging/model.py
|
Message._republish
|
def _republish(self):
"""
Re-publishes updated message
"""
mq_channel = self.channel._connect_mq()
mq_channel.basic_publish(exchange=self.channel.key, routing_key='',
body=json.dumps(self.serialize()))
|
python
|
def _republish(self):
"""
Re-publishes updated message
"""
mq_channel = self.channel._connect_mq()
mq_channel.basic_publish(exchange=self.channel.key, routing_key='',
body=json.dumps(self.serialize()))
|
[
"def",
"_republish",
"(",
"self",
")",
":",
"mq_channel",
"=",
"self",
".",
"channel",
".",
"_connect_mq",
"(",
")",
"mq_channel",
".",
"basic_publish",
"(",
"exchange",
"=",
"self",
".",
"channel",
".",
"key",
",",
"routing_key",
"=",
"''",
",",
"body",
"=",
"json",
".",
"dumps",
"(",
"self",
".",
"serialize",
"(",
")",
")",
")"
] |
Re-publishes updated message
|
[
"Re",
"-",
"publishes",
"updated",
"message"
] |
train
|
https://github.com/zetaops/zengine/blob/b5bc32d3b37bca799f8985be916f04528ac79e4a/zengine/messaging/model.py#L404-L410
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
defaultCrawlId
|
def defaultCrawlId():
"""
Provide a reasonable default crawl name using the user name and date
"""
timestamp = datetime.now().isoformat().replace(':', '_')
user = getuser()
return '_'.join(('crawl', user, timestamp))
|
python
|
def defaultCrawlId():
"""
Provide a reasonable default crawl name using the user name and date
"""
timestamp = datetime.now().isoformat().replace(':', '_')
user = getuser()
return '_'.join(('crawl', user, timestamp))
|
[
"def",
"defaultCrawlId",
"(",
")",
":",
"timestamp",
"=",
"datetime",
".",
"now",
"(",
")",
".",
"isoformat",
"(",
")",
".",
"replace",
"(",
"':'",
",",
"'_'",
")",
"user",
"=",
"getuser",
"(",
")",
"return",
"'_'",
".",
"join",
"(",
"(",
"'crawl'",
",",
"user",
",",
"timestamp",
")",
")"
] |
Provide a reasonable default crawl name using the user name and date
|
[
"Provide",
"a",
"reasonable",
"default",
"crawl",
"name",
"using",
"the",
"user",
"name",
"and",
"date"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L91-L98
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
main
|
def main(argv=None):
"""Run Nutch command using REST API."""
global Verbose, Mock
if argv is None:
argv = sys.argv
if len(argv) < 5: die('Bad args')
try:
opts, argv = getopt.getopt(argv[1:], 'hs:p:mv',
['help', 'server=', 'port=', 'mock', 'verbose'])
except getopt.GetoptError as err:
# print help information and exit:
print(err) # will print something like "option -a not recognized"
die()
serverEndpoint = DefaultServerEndpoint
# TODO: Fix this
for opt, val in opts:
if opt in ('-h', '--help'): echo2(USAGE); sys.exit()
elif opt in ('-s', '--server'): serverEndpoint = val
elif opt in ('-p', '--port'): serverEndpoint = 'http://localhost:%s' % val
elif opt in ('-m', '--mock'): Mock = 1
elif opt in ('-v', '--verbose'): Verbose = 1
else: die(USAGE)
cmd = argv[0]
crawlId = argv[1]
confId = argv[2]
urlDir = argv[3]
args = {}
if len(argv) > 4: args = eval(argv[4])
nt = Nutch(crawlId, confId, serverEndpoint, urlDir)
nt.Jobs().create(cmd, **args)
|
python
|
def main(argv=None):
"""Run Nutch command using REST API."""
global Verbose, Mock
if argv is None:
argv = sys.argv
if len(argv) < 5: die('Bad args')
try:
opts, argv = getopt.getopt(argv[1:], 'hs:p:mv',
['help', 'server=', 'port=', 'mock', 'verbose'])
except getopt.GetoptError as err:
# print help information and exit:
print(err) # will print something like "option -a not recognized"
die()
serverEndpoint = DefaultServerEndpoint
# TODO: Fix this
for opt, val in opts:
if opt in ('-h', '--help'): echo2(USAGE); sys.exit()
elif opt in ('-s', '--server'): serverEndpoint = val
elif opt in ('-p', '--port'): serverEndpoint = 'http://localhost:%s' % val
elif opt in ('-m', '--mock'): Mock = 1
elif opt in ('-v', '--verbose'): Verbose = 1
else: die(USAGE)
cmd = argv[0]
crawlId = argv[1]
confId = argv[2]
urlDir = argv[3]
args = {}
if len(argv) > 4: args = eval(argv[4])
nt = Nutch(crawlId, confId, serverEndpoint, urlDir)
nt.Jobs().create(cmd, **args)
|
[
"def",
"main",
"(",
"argv",
"=",
"None",
")",
":",
"global",
"Verbose",
",",
"Mock",
"if",
"argv",
"is",
"None",
":",
"argv",
"=",
"sys",
".",
"argv",
"if",
"len",
"(",
"argv",
")",
"<",
"5",
":",
"die",
"(",
"'Bad args'",
")",
"try",
":",
"opts",
",",
"argv",
"=",
"getopt",
".",
"getopt",
"(",
"argv",
"[",
"1",
":",
"]",
",",
"'hs:p:mv'",
",",
"[",
"'help'",
",",
"'server='",
",",
"'port='",
",",
"'mock'",
",",
"'verbose'",
"]",
")",
"except",
"getopt",
".",
"GetoptError",
"as",
"err",
":",
"# print help information and exit:",
"print",
"(",
"err",
")",
"# will print something like \"option -a not recognized\"",
"die",
"(",
")",
"serverEndpoint",
"=",
"DefaultServerEndpoint",
"# TODO: Fix this",
"for",
"opt",
",",
"val",
"in",
"opts",
":",
"if",
"opt",
"in",
"(",
"'-h'",
",",
"'--help'",
")",
":",
"echo2",
"(",
"USAGE",
")",
"sys",
".",
"exit",
"(",
")",
"elif",
"opt",
"in",
"(",
"'-s'",
",",
"'--server'",
")",
":",
"serverEndpoint",
"=",
"val",
"elif",
"opt",
"in",
"(",
"'-p'",
",",
"'--port'",
")",
":",
"serverEndpoint",
"=",
"'http://localhost:%s'",
"%",
"val",
"elif",
"opt",
"in",
"(",
"'-m'",
",",
"'--mock'",
")",
":",
"Mock",
"=",
"1",
"elif",
"opt",
"in",
"(",
"'-v'",
",",
"'--verbose'",
")",
":",
"Verbose",
"=",
"1",
"else",
":",
"die",
"(",
"USAGE",
")",
"cmd",
"=",
"argv",
"[",
"0",
"]",
"crawlId",
"=",
"argv",
"[",
"1",
"]",
"confId",
"=",
"argv",
"[",
"2",
"]",
"urlDir",
"=",
"argv",
"[",
"3",
"]",
"args",
"=",
"{",
"}",
"if",
"len",
"(",
"argv",
")",
">",
"4",
":",
"args",
"=",
"eval",
"(",
"argv",
"[",
"4",
"]",
")",
"nt",
"=",
"Nutch",
"(",
"crawlId",
",",
"confId",
",",
"serverEndpoint",
",",
"urlDir",
")",
"nt",
".",
"Jobs",
"(",
")",
".",
"create",
"(",
"cmd",
",",
"*",
"*",
"args",
")"
] |
Run Nutch command using REST API.
|
[
"Run",
"Nutch",
"command",
"using",
"REST",
"API",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L716-L749
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
Server.call
|
def call(self, verb, servicePath, data=None, headers=None, forceText=False, sendJson=True):
"""Call the Nutch Server, do some error checking, and return the response.
:param verb: One of nutch.RequestVerbs
:param servicePath: path component of URL to append to endpoint, e.g. '/config'
:param data: Data to attach to this request
:param headers: headers to attach to this request, default are JsonAcceptHeader
:param forceText: don't trust the response headers and just get the text
:param sendJson: Whether to treat attached data as JSON or not
"""
default_data = {} if sendJson else ""
data = data if data else default_data
headers = headers if headers else JsonAcceptHeader.copy()
if not sendJson:
headers.update(TextSendHeader)
if verb not in RequestVerbs:
die('Server call verb must be one of %s' % str(RequestVerbs.keys()))
if Verbose:
echo2("%s Endpoint:" % verb.upper(), servicePath)
echo2("%s Request data:" % verb.upper(), data)
echo2("%s Request headers:" % verb.upper(), headers)
verbFn = RequestVerbs[verb]
if sendJson:
resp = verbFn(self.serverEndpoint + servicePath, json=data, headers=headers)
else:
resp = verbFn(self.serverEndpoint + servicePath, data=data, headers=headers)
if Verbose:
echo2("Response headers:", resp.headers)
echo2("Response status:", resp.status_code)
if resp.status_code != 200:
if self.raiseErrors:
error = NutchException("Unexpected server response: %d" % resp.status_code)
error.status_code = resp.status_code
raise error
else:
warn('Nutch server returned status:', resp.status_code)
if forceText or 'content-type' not in resp.headers or resp.headers['content-type'] == 'text/plain':
if Verbose:
echo2("Response text:", resp.text)
return resp.text
content_type = resp.headers['content-type']
if content_type == 'application/json' and not forceText:
if Verbose:
echo2("Response JSON:", resp.json())
return resp.json()
else:
die('Did not understand server response: %s' % resp.headers)
|
python
|
def call(self, verb, servicePath, data=None, headers=None, forceText=False, sendJson=True):
"""Call the Nutch Server, do some error checking, and return the response.
:param verb: One of nutch.RequestVerbs
:param servicePath: path component of URL to append to endpoint, e.g. '/config'
:param data: Data to attach to this request
:param headers: headers to attach to this request, default are JsonAcceptHeader
:param forceText: don't trust the response headers and just get the text
:param sendJson: Whether to treat attached data as JSON or not
"""
default_data = {} if sendJson else ""
data = data if data else default_data
headers = headers if headers else JsonAcceptHeader.copy()
if not sendJson:
headers.update(TextSendHeader)
if verb not in RequestVerbs:
die('Server call verb must be one of %s' % str(RequestVerbs.keys()))
if Verbose:
echo2("%s Endpoint:" % verb.upper(), servicePath)
echo2("%s Request data:" % verb.upper(), data)
echo2("%s Request headers:" % verb.upper(), headers)
verbFn = RequestVerbs[verb]
if sendJson:
resp = verbFn(self.serverEndpoint + servicePath, json=data, headers=headers)
else:
resp = verbFn(self.serverEndpoint + servicePath, data=data, headers=headers)
if Verbose:
echo2("Response headers:", resp.headers)
echo2("Response status:", resp.status_code)
if resp.status_code != 200:
if self.raiseErrors:
error = NutchException("Unexpected server response: %d" % resp.status_code)
error.status_code = resp.status_code
raise error
else:
warn('Nutch server returned status:', resp.status_code)
if forceText or 'content-type' not in resp.headers or resp.headers['content-type'] == 'text/plain':
if Verbose:
echo2("Response text:", resp.text)
return resp.text
content_type = resp.headers['content-type']
if content_type == 'application/json' and not forceText:
if Verbose:
echo2("Response JSON:", resp.json())
return resp.json()
else:
die('Did not understand server response: %s' % resp.headers)
|
[
"def",
"call",
"(",
"self",
",",
"verb",
",",
"servicePath",
",",
"data",
"=",
"None",
",",
"headers",
"=",
"None",
",",
"forceText",
"=",
"False",
",",
"sendJson",
"=",
"True",
")",
":",
"default_data",
"=",
"{",
"}",
"if",
"sendJson",
"else",
"\"\"",
"data",
"=",
"data",
"if",
"data",
"else",
"default_data",
"headers",
"=",
"headers",
"if",
"headers",
"else",
"JsonAcceptHeader",
".",
"copy",
"(",
")",
"if",
"not",
"sendJson",
":",
"headers",
".",
"update",
"(",
"TextSendHeader",
")",
"if",
"verb",
"not",
"in",
"RequestVerbs",
":",
"die",
"(",
"'Server call verb must be one of %s'",
"%",
"str",
"(",
"RequestVerbs",
".",
"keys",
"(",
")",
")",
")",
"if",
"Verbose",
":",
"echo2",
"(",
"\"%s Endpoint:\"",
"%",
"verb",
".",
"upper",
"(",
")",
",",
"servicePath",
")",
"echo2",
"(",
"\"%s Request data:\"",
"%",
"verb",
".",
"upper",
"(",
")",
",",
"data",
")",
"echo2",
"(",
"\"%s Request headers:\"",
"%",
"verb",
".",
"upper",
"(",
")",
",",
"headers",
")",
"verbFn",
"=",
"RequestVerbs",
"[",
"verb",
"]",
"if",
"sendJson",
":",
"resp",
"=",
"verbFn",
"(",
"self",
".",
"serverEndpoint",
"+",
"servicePath",
",",
"json",
"=",
"data",
",",
"headers",
"=",
"headers",
")",
"else",
":",
"resp",
"=",
"verbFn",
"(",
"self",
".",
"serverEndpoint",
"+",
"servicePath",
",",
"data",
"=",
"data",
",",
"headers",
"=",
"headers",
")",
"if",
"Verbose",
":",
"echo2",
"(",
"\"Response headers:\"",
",",
"resp",
".",
"headers",
")",
"echo2",
"(",
"\"Response status:\"",
",",
"resp",
".",
"status_code",
")",
"if",
"resp",
".",
"status_code",
"!=",
"200",
":",
"if",
"self",
".",
"raiseErrors",
":",
"error",
"=",
"NutchException",
"(",
"\"Unexpected server response: %d\"",
"%",
"resp",
".",
"status_code",
")",
"error",
".",
"status_code",
"=",
"resp",
".",
"status_code",
"raise",
"error",
"else",
":",
"warn",
"(",
"'Nutch server returned status:'",
",",
"resp",
".",
"status_code",
")",
"if",
"forceText",
"or",
"'content-type'",
"not",
"in",
"resp",
".",
"headers",
"or",
"resp",
".",
"headers",
"[",
"'content-type'",
"]",
"==",
"'text/plain'",
":",
"if",
"Verbose",
":",
"echo2",
"(",
"\"Response text:\"",
",",
"resp",
".",
"text",
")",
"return",
"resp",
".",
"text",
"content_type",
"=",
"resp",
".",
"headers",
"[",
"'content-type'",
"]",
"if",
"content_type",
"==",
"'application/json'",
"and",
"not",
"forceText",
":",
"if",
"Verbose",
":",
"echo2",
"(",
"\"Response JSON:\"",
",",
"resp",
".",
"json",
"(",
")",
")",
"return",
"resp",
".",
"json",
"(",
")",
"else",
":",
"die",
"(",
"'Did not understand server response: %s'",
"%",
"resp",
".",
"headers",
")"
] |
Call the Nutch Server, do some error checking, and return the response.
:param verb: One of nutch.RequestVerbs
:param servicePath: path component of URL to append to endpoint, e.g. '/config'
:param data: Data to attach to this request
:param headers: headers to attach to this request, default are JsonAcceptHeader
:param forceText: don't trust the response headers and just get the text
:param sendJson: Whether to treat attached data as JSON or not
|
[
"Call",
"the",
"Nutch",
"Server",
"do",
"some",
"error",
"checking",
"and",
"return",
"the",
"response",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L117-L170
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
ConfigClient.create
|
def create(self, cid, configData):
"""
Create a new named (cid) configuration from a parameter dictionary (config_data).
"""
configArgs = {'configId': cid, 'params': configData, 'force': True}
cid = self.server.call('post', "/config/create", configArgs, forceText=True, headers=TextAcceptHeader)
new_config = Config(cid, self.server)
return new_config
|
python
|
def create(self, cid, configData):
"""
Create a new named (cid) configuration from a parameter dictionary (config_data).
"""
configArgs = {'configId': cid, 'params': configData, 'force': True}
cid = self.server.call('post', "/config/create", configArgs, forceText=True, headers=TextAcceptHeader)
new_config = Config(cid, self.server)
return new_config
|
[
"def",
"create",
"(",
"self",
",",
"cid",
",",
"configData",
")",
":",
"configArgs",
"=",
"{",
"'configId'",
":",
"cid",
",",
"'params'",
":",
"configData",
",",
"'force'",
":",
"True",
"}",
"cid",
"=",
"self",
".",
"server",
".",
"call",
"(",
"'post'",
",",
"\"/config/create\"",
",",
"configArgs",
",",
"forceText",
"=",
"True",
",",
"headers",
"=",
"TextAcceptHeader",
")",
"new_config",
"=",
"Config",
"(",
"cid",
",",
"self",
".",
"server",
")",
"return",
"new_config"
] |
Create a new named (cid) configuration from a parameter dictionary (config_data).
|
[
"Create",
"a",
"new",
"named",
"(",
"cid",
")",
"configuration",
"from",
"a",
"parameter",
"dictionary",
"(",
"config_data",
")",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L278-L285
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
JobClient.list
|
def list(self, allJobs=False):
"""
Return list of jobs at this endpoint.
Call get(allJobs=True) to see all jobs, not just the ones managed by this Client
"""
jobs = self.server.call('get', '/job')
return [Job(job['id'], self.server) for job in jobs if allJobs or self._job_owned(job)]
|
python
|
def list(self, allJobs=False):
"""
Return list of jobs at this endpoint.
Call get(allJobs=True) to see all jobs, not just the ones managed by this Client
"""
jobs = self.server.call('get', '/job')
return [Job(job['id'], self.server) for job in jobs if allJobs or self._job_owned(job)]
|
[
"def",
"list",
"(",
"self",
",",
"allJobs",
"=",
"False",
")",
":",
"jobs",
"=",
"self",
".",
"server",
".",
"call",
"(",
"'get'",
",",
"'/job'",
")",
"return",
"[",
"Job",
"(",
"job",
"[",
"'id'",
"]",
",",
"self",
".",
"server",
")",
"for",
"job",
"in",
"jobs",
"if",
"allJobs",
"or",
"self",
".",
"_job_owned",
"(",
"job",
")",
"]"
] |
Return list of jobs at this endpoint.
Call get(allJobs=True) to see all jobs, not just the ones managed by this Client
|
[
"Return",
"list",
"of",
"jobs",
"at",
"this",
"endpoint",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L337-L346
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
JobClient.create
|
def create(self, command, **args):
"""
Create a job given a command
:param command: Nutch command, one of nutch.LegalJobs
:param args: Additional arguments to pass to the job
:return: The created Job
"""
command = command.upper()
if command not in LegalJobs:
warn('Nutch command must be one of: %s' % ', '.join(LegalJobs))
else:
echo2('Starting %s job with args %s' % (command, str(args)))
parameters = self.parameters.copy()
parameters['type'] = command
parameters['crawlId'] = self.crawlId
parameters['confId'] = self.confId
parameters['args'].update(args)
job_info = self.server.call('post', "/job/create", parameters, JsonAcceptHeader)
job = Job(job_info['id'], self.server)
return job
|
python
|
def create(self, command, **args):
"""
Create a job given a command
:param command: Nutch command, one of nutch.LegalJobs
:param args: Additional arguments to pass to the job
:return: The created Job
"""
command = command.upper()
if command not in LegalJobs:
warn('Nutch command must be one of: %s' % ', '.join(LegalJobs))
else:
echo2('Starting %s job with args %s' % (command, str(args)))
parameters = self.parameters.copy()
parameters['type'] = command
parameters['crawlId'] = self.crawlId
parameters['confId'] = self.confId
parameters['args'].update(args)
job_info = self.server.call('post', "/job/create", parameters, JsonAcceptHeader)
job = Job(job_info['id'], self.server)
return job
|
[
"def",
"create",
"(",
"self",
",",
"command",
",",
"*",
"*",
"args",
")",
":",
"command",
"=",
"command",
".",
"upper",
"(",
")",
"if",
"command",
"not",
"in",
"LegalJobs",
":",
"warn",
"(",
"'Nutch command must be one of: %s'",
"%",
"', '",
".",
"join",
"(",
"LegalJobs",
")",
")",
"else",
":",
"echo2",
"(",
"'Starting %s job with args %s'",
"%",
"(",
"command",
",",
"str",
"(",
"args",
")",
")",
")",
"parameters",
"=",
"self",
".",
"parameters",
".",
"copy",
"(",
")",
"parameters",
"[",
"'type'",
"]",
"=",
"command",
"parameters",
"[",
"'crawlId'",
"]",
"=",
"self",
".",
"crawlId",
"parameters",
"[",
"'confId'",
"]",
"=",
"self",
".",
"confId",
"parameters",
"[",
"'args'",
"]",
".",
"update",
"(",
"args",
")",
"job_info",
"=",
"self",
".",
"server",
".",
"call",
"(",
"'post'",
",",
"\"/job/create\"",
",",
"parameters",
",",
"JsonAcceptHeader",
")",
"job",
"=",
"Job",
"(",
"job_info",
"[",
"'id'",
"]",
",",
"self",
".",
"server",
")",
"return",
"job"
] |
Create a job given a command
:param command: Nutch command, one of nutch.LegalJobs
:param args: Additional arguments to pass to the job
:return: The created Job
|
[
"Create",
"a",
"job",
"given",
"a",
"command",
":",
"param",
"command",
":",
"Nutch",
"command",
"one",
"of",
"nutch",
".",
"LegalJobs",
":",
"param",
"args",
":",
"Additional",
"arguments",
"to",
"pass",
"to",
"the",
"job",
":",
"return",
":",
"The",
"created",
"Job"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L348-L370
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
JobClient.inject
|
def inject(self, seed=None, urlDir=None, **args):
"""
:param seed: A Seed object (this or urlDir must be specified)
:param urlDir: The directory on the server containing the seed list (this or urlDir must be specified)
:param args: Extra arguments for the job
:return: a created Job object
"""
if seed:
if urlDir and urlDir != seed.seedPath:
raise NutchException("Can't specify both seed and urlDir")
urlDir = seed.seedPath
elif urlDir:
pass
else:
raise NutchException("Must specify seed or urlDir")
args['url_dir'] = urlDir
return self.create('INJECT', **args)
|
python
|
def inject(self, seed=None, urlDir=None, **args):
"""
:param seed: A Seed object (this or urlDir must be specified)
:param urlDir: The directory on the server containing the seed list (this or urlDir must be specified)
:param args: Extra arguments for the job
:return: a created Job object
"""
if seed:
if urlDir and urlDir != seed.seedPath:
raise NutchException("Can't specify both seed and urlDir")
urlDir = seed.seedPath
elif urlDir:
pass
else:
raise NutchException("Must specify seed or urlDir")
args['url_dir'] = urlDir
return self.create('INJECT', **args)
|
[
"def",
"inject",
"(",
"self",
",",
"seed",
"=",
"None",
",",
"urlDir",
"=",
"None",
",",
"*",
"*",
"args",
")",
":",
"if",
"seed",
":",
"if",
"urlDir",
"and",
"urlDir",
"!=",
"seed",
".",
"seedPath",
":",
"raise",
"NutchException",
"(",
"\"Can't specify both seed and urlDir\"",
")",
"urlDir",
"=",
"seed",
".",
"seedPath",
"elif",
"urlDir",
":",
"pass",
"else",
":",
"raise",
"NutchException",
"(",
"\"Must specify seed or urlDir\"",
")",
"args",
"[",
"'url_dir'",
"]",
"=",
"urlDir",
"return",
"self",
".",
"create",
"(",
"'INJECT'",
",",
"*",
"*",
"args",
")"
] |
:param seed: A Seed object (this or urlDir must be specified)
:param urlDir: The directory on the server containing the seed list (this or urlDir must be specified)
:param args: Extra arguments for the job
:return: a created Job object
|
[
":",
"param",
"seed",
":",
"A",
"Seed",
"object",
"(",
"this",
"or",
"urlDir",
"must",
"be",
"specified",
")",
":",
"param",
"urlDir",
":",
"The",
"directory",
"on",
"the",
"server",
"containing",
"the",
"seed",
"list",
"(",
"this",
"or",
"urlDir",
"must",
"be",
"specified",
")",
":",
"param",
"args",
":",
"Extra",
"arguments",
"for",
"the",
"job",
":",
"return",
":",
"a",
"created",
"Job",
"object"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L374-L391
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
SeedClient.create
|
def create(self, sid, seedList):
"""
Create a new named (sid) Seed from a list of seed URLs
:param sid: the name to assign to the new seed list
:param seedList: the list of seeds to use
:return: the created Seed object
"""
seedUrl = lambda uid, url: {"id": uid, "url": url}
if not isinstance(seedList,tuple):
seedList = (seedList,)
seedListData = {
"id": "12345",
"name": sid,
"seedUrls": [seedUrl(uid, url) for uid, url in enumerate(seedList)]
}
# As per resolution of https://issues.apache.org/jira/browse/NUTCH-2123
seedPath = self.server.call('post', "/seed/create", seedListData, TextAcceptHeader)
new_seed = Seed(sid, seedPath, self.server)
return new_seed
|
python
|
def create(self, sid, seedList):
"""
Create a new named (sid) Seed from a list of seed URLs
:param sid: the name to assign to the new seed list
:param seedList: the list of seeds to use
:return: the created Seed object
"""
seedUrl = lambda uid, url: {"id": uid, "url": url}
if not isinstance(seedList,tuple):
seedList = (seedList,)
seedListData = {
"id": "12345",
"name": sid,
"seedUrls": [seedUrl(uid, url) for uid, url in enumerate(seedList)]
}
# As per resolution of https://issues.apache.org/jira/browse/NUTCH-2123
seedPath = self.server.call('post', "/seed/create", seedListData, TextAcceptHeader)
new_seed = Seed(sid, seedPath, self.server)
return new_seed
|
[
"def",
"create",
"(",
"self",
",",
"sid",
",",
"seedList",
")",
":",
"seedUrl",
"=",
"lambda",
"uid",
",",
"url",
":",
"{",
"\"id\"",
":",
"uid",
",",
"\"url\"",
":",
"url",
"}",
"if",
"not",
"isinstance",
"(",
"seedList",
",",
"tuple",
")",
":",
"seedList",
"=",
"(",
"seedList",
",",
")",
"seedListData",
"=",
"{",
"\"id\"",
":",
"\"12345\"",
",",
"\"name\"",
":",
"sid",
",",
"\"seedUrls\"",
":",
"[",
"seedUrl",
"(",
"uid",
",",
"url",
")",
"for",
"uid",
",",
"url",
"in",
"enumerate",
"(",
"seedList",
")",
"]",
"}",
"# As per resolution of https://issues.apache.org/jira/browse/NUTCH-2123",
"seedPath",
"=",
"self",
".",
"server",
".",
"call",
"(",
"'post'",
",",
"\"/seed/create\"",
",",
"seedListData",
",",
"TextAcceptHeader",
")",
"new_seed",
"=",
"Seed",
"(",
"sid",
",",
"seedPath",
",",
"self",
".",
"server",
")",
"return",
"new_seed"
] |
Create a new named (sid) Seed from a list of seed URLs
:param sid: the name to assign to the new seed list
:param seedList: the list of seeds to use
:return: the created Seed object
|
[
"Create",
"a",
"new",
"named",
"(",
"sid",
")",
"Seed",
"from",
"a",
"list",
"of",
"seed",
"URLs"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L419-L442
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
SeedClient.createFromFile
|
def createFromFile(self, sid, filename):
"""
Create a new named (sid) Seed from a file containing URLs
It's assumed URLs are whitespace seperated.
:param sid: the name to assign to the new seed list
:param filename: the name of the file that contains URLs
:return: the created Seed object
"""
urls = []
with open(filename) as f:
for line in f:
for url in line.split():
urls.append(url)
return self.create(sid, tuple(urls))
|
python
|
def createFromFile(self, sid, filename):
"""
Create a new named (sid) Seed from a file containing URLs
It's assumed URLs are whitespace seperated.
:param sid: the name to assign to the new seed list
:param filename: the name of the file that contains URLs
:return: the created Seed object
"""
urls = []
with open(filename) as f:
for line in f:
for url in line.split():
urls.append(url)
return self.create(sid, tuple(urls))
|
[
"def",
"createFromFile",
"(",
"self",
",",
"sid",
",",
"filename",
")",
":",
"urls",
"=",
"[",
"]",
"with",
"open",
"(",
"filename",
")",
"as",
"f",
":",
"for",
"line",
"in",
"f",
":",
"for",
"url",
"in",
"line",
".",
"split",
"(",
")",
":",
"urls",
".",
"append",
"(",
"url",
")",
"return",
"self",
".",
"create",
"(",
"sid",
",",
"tuple",
"(",
"urls",
")",
")"
] |
Create a new named (sid) Seed from a file containing URLs
It's assumed URLs are whitespace seperated.
:param sid: the name to assign to the new seed list
:param filename: the name of the file that contains URLs
:return: the created Seed object
|
[
"Create",
"a",
"new",
"named",
"(",
"sid",
")",
"Seed",
"from",
"a",
"file",
"containing",
"URLs",
"It",
"s",
"assumed",
"URLs",
"are",
"whitespace",
"seperated",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L444-L460
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
CrawlClient._nextJob
|
def _nextJob(self, job, nextRound=True):
"""
Given a completed job, start the next job in the round, or return None
:param nextRound: whether to start jobs from the next round if the current round is completed.
:return: the newly started Job, or None if no job was started
"""
jobInfo = job.info()
assert jobInfo['state'] == 'FINISHED'
roundEnd = False
if jobInfo['type'] == 'INJECT':
nextCommand = 'GENERATE'
elif jobInfo['type'] == 'GENERATE':
nextCommand = 'FETCH'
elif jobInfo['type'] == 'FETCH':
nextCommand = 'PARSE'
elif jobInfo['type'] == 'PARSE':
nextCommand = 'UPDATEDB'
elif jobInfo['type'] == 'UPDATEDB':
nextCommand = 'INVERTLINKS'
elif jobInfo['type'] == 'INVERTLINKS':
nextCommand = 'DEDUP'
elif jobInfo['type'] == 'DEDUP':
if self.enable_index:
nextCommand = 'INDEX'
else:
roundEnd = True
elif jobInfo['type'] == 'INDEX':
roundEnd = True
else:
raise NutchException("Unrecognized job type {}".format(jobInfo['type']))
if roundEnd:
if nextRound and self.currentRound < self.totalRounds:
nextCommand = 'GENERATE'
self.currentRound += 1
else:
return None
return self.jobClient.create(nextCommand)
|
python
|
def _nextJob(self, job, nextRound=True):
"""
Given a completed job, start the next job in the round, or return None
:param nextRound: whether to start jobs from the next round if the current round is completed.
:return: the newly started Job, or None if no job was started
"""
jobInfo = job.info()
assert jobInfo['state'] == 'FINISHED'
roundEnd = False
if jobInfo['type'] == 'INJECT':
nextCommand = 'GENERATE'
elif jobInfo['type'] == 'GENERATE':
nextCommand = 'FETCH'
elif jobInfo['type'] == 'FETCH':
nextCommand = 'PARSE'
elif jobInfo['type'] == 'PARSE':
nextCommand = 'UPDATEDB'
elif jobInfo['type'] == 'UPDATEDB':
nextCommand = 'INVERTLINKS'
elif jobInfo['type'] == 'INVERTLINKS':
nextCommand = 'DEDUP'
elif jobInfo['type'] == 'DEDUP':
if self.enable_index:
nextCommand = 'INDEX'
else:
roundEnd = True
elif jobInfo['type'] == 'INDEX':
roundEnd = True
else:
raise NutchException("Unrecognized job type {}".format(jobInfo['type']))
if roundEnd:
if nextRound and self.currentRound < self.totalRounds:
nextCommand = 'GENERATE'
self.currentRound += 1
else:
return None
return self.jobClient.create(nextCommand)
|
[
"def",
"_nextJob",
"(",
"self",
",",
"job",
",",
"nextRound",
"=",
"True",
")",
":",
"jobInfo",
"=",
"job",
".",
"info",
"(",
")",
"assert",
"jobInfo",
"[",
"'state'",
"]",
"==",
"'FINISHED'",
"roundEnd",
"=",
"False",
"if",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'INJECT'",
":",
"nextCommand",
"=",
"'GENERATE'",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'GENERATE'",
":",
"nextCommand",
"=",
"'FETCH'",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'FETCH'",
":",
"nextCommand",
"=",
"'PARSE'",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'PARSE'",
":",
"nextCommand",
"=",
"'UPDATEDB'",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'UPDATEDB'",
":",
"nextCommand",
"=",
"'INVERTLINKS'",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'INVERTLINKS'",
":",
"nextCommand",
"=",
"'DEDUP'",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'DEDUP'",
":",
"if",
"self",
".",
"enable_index",
":",
"nextCommand",
"=",
"'INDEX'",
"else",
":",
"roundEnd",
"=",
"True",
"elif",
"jobInfo",
"[",
"'type'",
"]",
"==",
"'INDEX'",
":",
"roundEnd",
"=",
"True",
"else",
":",
"raise",
"NutchException",
"(",
"\"Unrecognized job type {}\"",
".",
"format",
"(",
"jobInfo",
"[",
"'type'",
"]",
")",
")",
"if",
"roundEnd",
":",
"if",
"nextRound",
"and",
"self",
".",
"currentRound",
"<",
"self",
".",
"totalRounds",
":",
"nextCommand",
"=",
"'GENERATE'",
"self",
".",
"currentRound",
"+=",
"1",
"else",
":",
"return",
"None",
"return",
"self",
".",
"jobClient",
".",
"create",
"(",
"nextCommand",
")"
] |
Given a completed job, start the next job in the round, or return None
:param nextRound: whether to start jobs from the next round if the current round is completed.
:return: the newly started Job, or None if no job was started
|
[
"Given",
"a",
"completed",
"job",
"start",
"the",
"next",
"job",
"in",
"the",
"round",
"or",
"return",
"None"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L492-L533
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
CrawlClient.progress
|
def progress(self, nextRound=True):
"""
Check the status of the current job, activate the next job if it's finished, and return the active job
If the current job has failed, a NutchCrawlException will be raised with no jobs attached.
:param nextRound: whether to start jobs from the next round if the current job/round is completed.
:return: the currently running Job, or None if no jobs are running.
"""
currentJob = self.currentJob
if currentJob is None:
return currentJob
jobInfo = currentJob.info()
if jobInfo['state'] == 'RUNNING':
return currentJob
elif jobInfo['state'] == 'FINISHED':
nextJob = self._nextJob(currentJob, nextRound)
self.currentJob = nextJob
return nextJob
else:
error = NutchCrawlException("Unexpected job state: {}".format(jobInfo['state']))
error.current_job = currentJob
raise NutchCrawlException
|
python
|
def progress(self, nextRound=True):
"""
Check the status of the current job, activate the next job if it's finished, and return the active job
If the current job has failed, a NutchCrawlException will be raised with no jobs attached.
:param nextRound: whether to start jobs from the next round if the current job/round is completed.
:return: the currently running Job, or None if no jobs are running.
"""
currentJob = self.currentJob
if currentJob is None:
return currentJob
jobInfo = currentJob.info()
if jobInfo['state'] == 'RUNNING':
return currentJob
elif jobInfo['state'] == 'FINISHED':
nextJob = self._nextJob(currentJob, nextRound)
self.currentJob = nextJob
return nextJob
else:
error = NutchCrawlException("Unexpected job state: {}".format(jobInfo['state']))
error.current_job = currentJob
raise NutchCrawlException
|
[
"def",
"progress",
"(",
"self",
",",
"nextRound",
"=",
"True",
")",
":",
"currentJob",
"=",
"self",
".",
"currentJob",
"if",
"currentJob",
"is",
"None",
":",
"return",
"currentJob",
"jobInfo",
"=",
"currentJob",
".",
"info",
"(",
")",
"if",
"jobInfo",
"[",
"'state'",
"]",
"==",
"'RUNNING'",
":",
"return",
"currentJob",
"elif",
"jobInfo",
"[",
"'state'",
"]",
"==",
"'FINISHED'",
":",
"nextJob",
"=",
"self",
".",
"_nextJob",
"(",
"currentJob",
",",
"nextRound",
")",
"self",
".",
"currentJob",
"=",
"nextJob",
"return",
"nextJob",
"else",
":",
"error",
"=",
"NutchCrawlException",
"(",
"\"Unexpected job state: {}\"",
".",
"format",
"(",
"jobInfo",
"[",
"'state'",
"]",
")",
")",
"error",
".",
"current_job",
"=",
"currentJob",
"raise",
"NutchCrawlException"
] |
Check the status of the current job, activate the next job if it's finished, and return the active job
If the current job has failed, a NutchCrawlException will be raised with no jobs attached.
:param nextRound: whether to start jobs from the next round if the current job/round is completed.
:return: the currently running Job, or None if no jobs are running.
|
[
"Check",
"the",
"status",
"of",
"the",
"current",
"job",
"activate",
"the",
"next",
"job",
"if",
"it",
"s",
"finished",
"and",
"return",
"the",
"active",
"job"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L535-L560
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
CrawlClient.nextRound
|
def nextRound(self):
"""
Execute all jobs in the current round and return when they have finished.
If a job fails, a NutchCrawlException will be raised, with all completed jobs from this round attached
to the exception.
:return: a list of all completed Jobs
"""
finishedJobs = []
if self.currentJob is None:
self.currentJob = self.jobClient.create('GENERATE')
activeJob = self.progress(nextRound=False)
while activeJob:
oldJob = activeJob
activeJob = self.progress(nextRound=False) # updates self.currentJob
if oldJob and oldJob != activeJob:
finishedJobs.append(oldJob)
sleep(self.sleepTime)
self.currentRound += 1
return finishedJobs
|
python
|
def nextRound(self):
"""
Execute all jobs in the current round and return when they have finished.
If a job fails, a NutchCrawlException will be raised, with all completed jobs from this round attached
to the exception.
:return: a list of all completed Jobs
"""
finishedJobs = []
if self.currentJob is None:
self.currentJob = self.jobClient.create('GENERATE')
activeJob = self.progress(nextRound=False)
while activeJob:
oldJob = activeJob
activeJob = self.progress(nextRound=False) # updates self.currentJob
if oldJob and oldJob != activeJob:
finishedJobs.append(oldJob)
sleep(self.sleepTime)
self.currentRound += 1
return finishedJobs
|
[
"def",
"nextRound",
"(",
"self",
")",
":",
"finishedJobs",
"=",
"[",
"]",
"if",
"self",
".",
"currentJob",
"is",
"None",
":",
"self",
".",
"currentJob",
"=",
"self",
".",
"jobClient",
".",
"create",
"(",
"'GENERATE'",
")",
"activeJob",
"=",
"self",
".",
"progress",
"(",
"nextRound",
"=",
"False",
")",
"while",
"activeJob",
":",
"oldJob",
"=",
"activeJob",
"activeJob",
"=",
"self",
".",
"progress",
"(",
"nextRound",
"=",
"False",
")",
"# updates self.currentJob",
"if",
"oldJob",
"and",
"oldJob",
"!=",
"activeJob",
":",
"finishedJobs",
".",
"append",
"(",
"oldJob",
")",
"sleep",
"(",
"self",
".",
"sleepTime",
")",
"self",
".",
"currentRound",
"+=",
"1",
"return",
"finishedJobs"
] |
Execute all jobs in the current round and return when they have finished.
If a job fails, a NutchCrawlException will be raised, with all completed jobs from this round attached
to the exception.
:return: a list of all completed Jobs
|
[
"Execute",
"all",
"jobs",
"in",
"the",
"current",
"round",
"and",
"return",
"when",
"they",
"have",
"finished",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L573-L595
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
CrawlClient.waitAll
|
def waitAll(self):
"""
Execute all queued rounds and return when they have finished.
If a job fails, a NutchCrawlException will be raised, with all completed jobs attached
to the exception
:return: a list of jobs completed for each round, organized by round (list-of-lists)
"""
finishedRounds = [self.nextRound()]
while self.currentRound < self.totalRounds:
finishedRounds.append(self.nextRound())
return finishedRounds
|
python
|
def waitAll(self):
"""
Execute all queued rounds and return when they have finished.
If a job fails, a NutchCrawlException will be raised, with all completed jobs attached
to the exception
:return: a list of jobs completed for each round, organized by round (list-of-lists)
"""
finishedRounds = [self.nextRound()]
while self.currentRound < self.totalRounds:
finishedRounds.append(self.nextRound())
return finishedRounds
|
[
"def",
"waitAll",
"(",
"self",
")",
":",
"finishedRounds",
"=",
"[",
"self",
".",
"nextRound",
"(",
")",
"]",
"while",
"self",
".",
"currentRound",
"<",
"self",
".",
"totalRounds",
":",
"finishedRounds",
".",
"append",
"(",
"self",
".",
"nextRound",
"(",
")",
")",
"return",
"finishedRounds"
] |
Execute all queued rounds and return when they have finished.
If a job fails, a NutchCrawlException will be raised, with all completed jobs attached
to the exception
:return: a list of jobs completed for each round, organized by round (list-of-lists)
|
[
"Execute",
"all",
"queued",
"rounds",
"and",
"return",
"when",
"they",
"have",
"finished",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L597-L612
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
Nutch.Jobs
|
def Jobs(self, crawlId=None):
"""
Create a JobClient for listing and creating jobs.
The JobClient inherits the confId from the Nutch client.
:param crawlId: crawlIds to use for this client. If not provided, will be generated
by nutch.defaultCrawlId()
:return: a JobClient
"""
crawlId = crawlId if crawlId else defaultCrawlId()
return JobClient(self.server, crawlId, self.confId)
|
python
|
def Jobs(self, crawlId=None):
"""
Create a JobClient for listing and creating jobs.
The JobClient inherits the confId from the Nutch client.
:param crawlId: crawlIds to use for this client. If not provided, will be generated
by nutch.defaultCrawlId()
:return: a JobClient
"""
crawlId = crawlId if crawlId else defaultCrawlId()
return JobClient(self.server, crawlId, self.confId)
|
[
"def",
"Jobs",
"(",
"self",
",",
"crawlId",
"=",
"None",
")",
":",
"crawlId",
"=",
"crawlId",
"if",
"crawlId",
"else",
"defaultCrawlId",
"(",
")",
"return",
"JobClient",
"(",
"self",
".",
"server",
",",
"crawlId",
",",
"self",
".",
"confId",
")"
] |
Create a JobClient for listing and creating jobs.
The JobClient inherits the confId from the Nutch client.
:param crawlId: crawlIds to use for this client. If not provided, will be generated
by nutch.defaultCrawlId()
:return: a JobClient
|
[
"Create",
"a",
"JobClient",
"for",
"listing",
"and",
"creating",
"jobs",
".",
"The",
"JobClient",
"inherits",
"the",
"confId",
"from",
"the",
"Nutch",
"client",
"."
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L656-L666
|
chrismattmann/nutch-python
|
nutch/nutch.py
|
Nutch.Crawl
|
def Crawl(self, seed, seedClient=None, jobClient=None, rounds=1, index=True):
"""
Launch a crawl using the given seed
:param seed: Type (Seed or SeedList) - used for crawl
:param seedClient: if a SeedList is given, the SeedClient to upload, if None a default will be created
:param jobClient: the JobClient to be used, if None a default will be created
:param rounds: the number of rounds in the crawl
:return: a CrawlClient to monitor and control the crawl
"""
if seedClient is None:
seedClient = self.Seeds()
if jobClient is None:
jobClient = self.Jobs()
if type(seed) != Seed:
seed = seedClient.create(jobClient.crawlId + '_seeds', seed)
return CrawlClient(self.server, seed, jobClient, rounds, index)
|
python
|
def Crawl(self, seed, seedClient=None, jobClient=None, rounds=1, index=True):
"""
Launch a crawl using the given seed
:param seed: Type (Seed or SeedList) - used for crawl
:param seedClient: if a SeedList is given, the SeedClient to upload, if None a default will be created
:param jobClient: the JobClient to be used, if None a default will be created
:param rounds: the number of rounds in the crawl
:return: a CrawlClient to monitor and control the crawl
"""
if seedClient is None:
seedClient = self.Seeds()
if jobClient is None:
jobClient = self.Jobs()
if type(seed) != Seed:
seed = seedClient.create(jobClient.crawlId + '_seeds', seed)
return CrawlClient(self.server, seed, jobClient, rounds, index)
|
[
"def",
"Crawl",
"(",
"self",
",",
"seed",
",",
"seedClient",
"=",
"None",
",",
"jobClient",
"=",
"None",
",",
"rounds",
"=",
"1",
",",
"index",
"=",
"True",
")",
":",
"if",
"seedClient",
"is",
"None",
":",
"seedClient",
"=",
"self",
".",
"Seeds",
"(",
")",
"if",
"jobClient",
"is",
"None",
":",
"jobClient",
"=",
"self",
".",
"Jobs",
"(",
")",
"if",
"type",
"(",
"seed",
")",
"!=",
"Seed",
":",
"seed",
"=",
"seedClient",
".",
"create",
"(",
"jobClient",
".",
"crawlId",
"+",
"'_seeds'",
",",
"seed",
")",
"return",
"CrawlClient",
"(",
"self",
".",
"server",
",",
"seed",
",",
"jobClient",
",",
"rounds",
",",
"index",
")"
] |
Launch a crawl using the given seed
:param seed: Type (Seed or SeedList) - used for crawl
:param seedClient: if a SeedList is given, the SeedClient to upload, if None a default will be created
:param jobClient: the JobClient to be used, if None a default will be created
:param rounds: the number of rounds in the crawl
:return: a CrawlClient to monitor and control the crawl
|
[
"Launch",
"a",
"crawl",
"using",
"the",
"given",
"seed",
":",
"param",
"seed",
":",
"Type",
"(",
"Seed",
"or",
"SeedList",
")",
"-",
"used",
"for",
"crawl",
":",
"param",
"seedClient",
":",
"if",
"a",
"SeedList",
"is",
"given",
"the",
"SeedClient",
"to",
"upload",
"if",
"None",
"a",
"default",
"will",
"be",
"created",
":",
"param",
"jobClient",
":",
"the",
"JobClient",
"to",
"be",
"used",
"if",
"None",
"a",
"default",
"will",
"be",
"created",
":",
"param",
"rounds",
":",
"the",
"number",
"of",
"rounds",
"in",
"the",
"crawl",
":",
"return",
":",
"a",
"CrawlClient",
"to",
"monitor",
"and",
"control",
"the",
"crawl"
] |
train
|
https://github.com/chrismattmann/nutch-python/blob/07ae182e283b2f74ef062ddfa20a690a59ab6f5a/nutch/nutch.py#L677-L693
|
deep-compute/logagg
|
logagg/formatters.py
|
haproxy
|
def haproxy(line):
#TODO Handle all message formats
'''
>>> import pprint
>>> input_line1 = 'Apr 24 00:00:02 node haproxy[12298]: 1.1.1.1:48660 [24/Apr/2019:00:00:02.358] pre-staging~ pre-staging_doc/pre-staging_active 261/0/2/8/271 200 2406 - - ---- 4/4/0/1/0 0/0 {AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018} "GET /doc/api/get?call=apple HTTP/1.1"'
>>> output_line1 = haproxy(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'Tc': 2.0,
'Tq': 261.0,
'Tr': 8.0,
'Tw': 0.0,
'_api': '/doc/api/get?call=apple',
'_headers': ['AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018'],
'actconn': 4,
'backend': 'pre-staging_doc/pre-staging_active',
'backend_queue': 0,
'beconn': 1,
'bytes_read': 2406.0,
'client_port': '48660',
'client_server': '1.1.1.1',
'feconn': 4,
'front_end': 'pre-staging~',
'haproxy_server': 'node',
'method': 'GET',
'resp_time': 271.0,
'retries': 0,
'srv_conn': 0,
'srv_queue': 0,
'status': '200',
'timestamp': '2019-04-24T00:00:02.358000'},
'event': 'haproxy_event',
'timestamp': '2019-04-24T00:00:02.358000',
'type': 'metric'}
'''
_line = line.strip().split()
log = {}
log['client_server'] = _line[5].split(':')[0].strip()
log['client_port'] = _line[5].split(':')[1].strip()
_timestamp = re.findall(r'\[(.*?)\]', _line[6])[0]
log['timestamp'] = datetime.datetime.strptime(_timestamp, '%d/%b/%Y:%H:%M:%S.%f').isoformat()
log['front_end'] = _line[7].strip()
log['backend'] = _line[8].strip()
log['Tq'] = float(_line[9].split('/')[0].strip())
log['Tw'] = float(_line[9].split('/')[1].strip())
log['Tc'] = float(_line[9].split('/')[2].strip())
log['Tr'] = float(_line[9].split('/')[3].strip())
log['resp_time'] = float(_line[9].split('/')[-1].strip())
log['status'] = _line[10].strip()
log['bytes_read'] = float(_line[11].strip())
log['_headers'] = re.findall(r'{(.*)}', line)
log['haproxy_server'] = _line[3].strip()
log['method'] = _line[-3].strip('"').strip()
log['_api'] = _line[-2].strip()
log['retries'] = int(_line[15].split('/')[-1].strip())
log['actconn'] = int(_line[15].split('/')[0].strip())
log['feconn'] = int(_line[15].split('/')[1].strip())
log['beconn'] = int(_line[15].split('/')[-2].strip())
log['srv_conn'] = int(_line[15].split('/')[-3].strip())
log['srv_queue'] = int(_line[16].split('/')[0].strip())
log['backend_queue'] = int(_line[16].split('/')[1].strip())
return dict(
data=log,
event='haproxy_event',
timestamp=log.get('timestamp'),
type='metric'
)
|
python
|
def haproxy(line):
#TODO Handle all message formats
'''
>>> import pprint
>>> input_line1 = 'Apr 24 00:00:02 node haproxy[12298]: 1.1.1.1:48660 [24/Apr/2019:00:00:02.358] pre-staging~ pre-staging_doc/pre-staging_active 261/0/2/8/271 200 2406 - - ---- 4/4/0/1/0 0/0 {AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018} "GET /doc/api/get?call=apple HTTP/1.1"'
>>> output_line1 = haproxy(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'Tc': 2.0,
'Tq': 261.0,
'Tr': 8.0,
'Tw': 0.0,
'_api': '/doc/api/get?call=apple',
'_headers': ['AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018'],
'actconn': 4,
'backend': 'pre-staging_doc/pre-staging_active',
'backend_queue': 0,
'beconn': 1,
'bytes_read': 2406.0,
'client_port': '48660',
'client_server': '1.1.1.1',
'feconn': 4,
'front_end': 'pre-staging~',
'haproxy_server': 'node',
'method': 'GET',
'resp_time': 271.0,
'retries': 0,
'srv_conn': 0,
'srv_queue': 0,
'status': '200',
'timestamp': '2019-04-24T00:00:02.358000'},
'event': 'haproxy_event',
'timestamp': '2019-04-24T00:00:02.358000',
'type': 'metric'}
'''
_line = line.strip().split()
log = {}
log['client_server'] = _line[5].split(':')[0].strip()
log['client_port'] = _line[5].split(':')[1].strip()
_timestamp = re.findall(r'\[(.*?)\]', _line[6])[0]
log['timestamp'] = datetime.datetime.strptime(_timestamp, '%d/%b/%Y:%H:%M:%S.%f').isoformat()
log['front_end'] = _line[7].strip()
log['backend'] = _line[8].strip()
log['Tq'] = float(_line[9].split('/')[0].strip())
log['Tw'] = float(_line[9].split('/')[1].strip())
log['Tc'] = float(_line[9].split('/')[2].strip())
log['Tr'] = float(_line[9].split('/')[3].strip())
log['resp_time'] = float(_line[9].split('/')[-1].strip())
log['status'] = _line[10].strip()
log['bytes_read'] = float(_line[11].strip())
log['_headers'] = re.findall(r'{(.*)}', line)
log['haproxy_server'] = _line[3].strip()
log['method'] = _line[-3].strip('"').strip()
log['_api'] = _line[-2].strip()
log['retries'] = int(_line[15].split('/')[-1].strip())
log['actconn'] = int(_line[15].split('/')[0].strip())
log['feconn'] = int(_line[15].split('/')[1].strip())
log['beconn'] = int(_line[15].split('/')[-2].strip())
log['srv_conn'] = int(_line[15].split('/')[-3].strip())
log['srv_queue'] = int(_line[16].split('/')[0].strip())
log['backend_queue'] = int(_line[16].split('/')[1].strip())
return dict(
data=log,
event='haproxy_event',
timestamp=log.get('timestamp'),
type='metric'
)
|
[
"def",
"haproxy",
"(",
"line",
")",
":",
"#TODO Handle all message formats",
"_line",
"=",
"line",
".",
"strip",
"(",
")",
".",
"split",
"(",
")",
"log",
"=",
"{",
"}",
"log",
"[",
"'client_server'",
"]",
"=",
"_line",
"[",
"5",
"]",
".",
"split",
"(",
"':'",
")",
"[",
"0",
"]",
".",
"strip",
"(",
")",
"log",
"[",
"'client_port'",
"]",
"=",
"_line",
"[",
"5",
"]",
".",
"split",
"(",
"':'",
")",
"[",
"1",
"]",
".",
"strip",
"(",
")",
"_timestamp",
"=",
"re",
".",
"findall",
"(",
"r'\\[(.*?)\\]'",
",",
"_line",
"[",
"6",
"]",
")",
"[",
"0",
"]",
"log",
"[",
"'timestamp'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"strptime",
"(",
"_timestamp",
",",
"'%d/%b/%Y:%H:%M:%S.%f'",
")",
".",
"isoformat",
"(",
")",
"log",
"[",
"'front_end'",
"]",
"=",
"_line",
"[",
"7",
"]",
".",
"strip",
"(",
")",
"log",
"[",
"'backend'",
"]",
"=",
"_line",
"[",
"8",
"]",
".",
"strip",
"(",
")",
"log",
"[",
"'Tq'",
"]",
"=",
"float",
"(",
"_line",
"[",
"9",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"0",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'Tw'",
"]",
"=",
"float",
"(",
"_line",
"[",
"9",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"1",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'Tc'",
"]",
"=",
"float",
"(",
"_line",
"[",
"9",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"2",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'Tr'",
"]",
"=",
"float",
"(",
"_line",
"[",
"9",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"3",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'resp_time'",
"]",
"=",
"float",
"(",
"_line",
"[",
"9",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"-",
"1",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'status'",
"]",
"=",
"_line",
"[",
"10",
"]",
".",
"strip",
"(",
")",
"log",
"[",
"'bytes_read'",
"]",
"=",
"float",
"(",
"_line",
"[",
"11",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'_headers'",
"]",
"=",
"re",
".",
"findall",
"(",
"r'{(.*)}'",
",",
"line",
")",
"log",
"[",
"'haproxy_server'",
"]",
"=",
"_line",
"[",
"3",
"]",
".",
"strip",
"(",
")",
"log",
"[",
"'method'",
"]",
"=",
"_line",
"[",
"-",
"3",
"]",
".",
"strip",
"(",
"'\"'",
")",
".",
"strip",
"(",
")",
"log",
"[",
"'_api'",
"]",
"=",
"_line",
"[",
"-",
"2",
"]",
".",
"strip",
"(",
")",
"log",
"[",
"'retries'",
"]",
"=",
"int",
"(",
"_line",
"[",
"15",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"-",
"1",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'actconn'",
"]",
"=",
"int",
"(",
"_line",
"[",
"15",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"0",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'feconn'",
"]",
"=",
"int",
"(",
"_line",
"[",
"15",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"1",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'beconn'",
"]",
"=",
"int",
"(",
"_line",
"[",
"15",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"-",
"2",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'srv_conn'",
"]",
"=",
"int",
"(",
"_line",
"[",
"15",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"-",
"3",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'srv_queue'",
"]",
"=",
"int",
"(",
"_line",
"[",
"16",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"0",
"]",
".",
"strip",
"(",
")",
")",
"log",
"[",
"'backend_queue'",
"]",
"=",
"int",
"(",
"_line",
"[",
"16",
"]",
".",
"split",
"(",
"'/'",
")",
"[",
"1",
"]",
".",
"strip",
"(",
")",
")",
"return",
"dict",
"(",
"data",
"=",
"log",
",",
"event",
"=",
"'haproxy_event'",
",",
"timestamp",
"=",
"log",
".",
"get",
"(",
"'timestamp'",
")",
",",
"type",
"=",
"'metric'",
")"
] |
>>> import pprint
>>> input_line1 = 'Apr 24 00:00:02 node haproxy[12298]: 1.1.1.1:48660 [24/Apr/2019:00:00:02.358] pre-staging~ pre-staging_doc/pre-staging_active 261/0/2/8/271 200 2406 - - ---- 4/4/0/1/0 0/0 {AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018} "GET /doc/api/get?call=apple HTTP/1.1"'
>>> output_line1 = haproxy(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'Tc': 2.0,
'Tq': 261.0,
'Tr': 8.0,
'Tw': 0.0,
'_api': '/doc/api/get?call=apple',
'_headers': ['AAAAAA:AAAAA_AAAAA:AAAAA_AAAAA_AAAAA:300A||| user@mail.net:sdasdasdasdsdasAHDivsjd=|user@mail.net|2018'],
'actconn': 4,
'backend': 'pre-staging_doc/pre-staging_active',
'backend_queue': 0,
'beconn': 1,
'bytes_read': 2406.0,
'client_port': '48660',
'client_server': '1.1.1.1',
'feconn': 4,
'front_end': 'pre-staging~',
'haproxy_server': 'node',
'method': 'GET',
'resp_time': 271.0,
'retries': 0,
'srv_conn': 0,
'srv_queue': 0,
'status': '200',
'timestamp': '2019-04-24T00:00:02.358000'},
'event': 'haproxy_event',
'timestamp': '2019-04-24T00:00:02.358000',
'type': 'metric'}
|
[
">>>",
"import",
"pprint",
">>>",
"input_line1",
"=",
"Apr",
"24",
"00",
":",
"00",
":",
"02",
"node",
"haproxy",
"[",
"12298",
"]",
":",
"1",
".",
"1",
".",
"1",
".",
"1",
":",
"48660",
"[",
"24",
"/",
"Apr",
"/",
"2019",
":",
"00",
":",
"00",
":",
"02",
".",
"358",
"]",
"pre",
"-",
"staging~",
"pre",
"-",
"staging_doc",
"/",
"pre",
"-",
"staging_active",
"261",
"/",
"0",
"/",
"2",
"/",
"8",
"/",
"271",
"200",
"2406",
"-",
"-",
"----",
"4",
"/",
"4",
"/",
"0",
"/",
"1",
"/",
"0",
"0",
"/",
"0",
"{",
"AAAAAA",
":",
"AAAAA_AAAAA",
":",
"AAAAA_AAAAA_AAAAA",
":",
"300A|||",
"user"
] |
train
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L19-L94
|
deep-compute/logagg
|
logagg/formatters.py
|
nginx_access
|
def nginx_access(line):
'''
>>> import pprint
>>> input_line1 = '{ \
"remote_addr": "127.0.0.1","remote_user": "-","timestamp": "1515144699.201", \
"request": "GET / HTTP/1.1","status": "200","request_time": "0.000", \
"body_bytes_sent": "396","http_referer": "-","http_user_agent": "python-requests/2.18.4", \
"http_x_forwarded_for": "-","upstream_response_time": "-" \
}'
>>> output_line1 = nginx_access(input_line1)
>>> pprint.pprint(output_line1)
{'data': {u'body_bytes_sent': 396.0,
u'http_referer': u'-',
u'http_user_agent': u'python-requests/2.18.4',
u'http_x_forwarded_for': u'-',
u'remote_addr': u'127.0.0.1',
u'remote_user': u'-',
u'request': u'GET / HTTP/1.1',
u'request_time': 0.0,
u'status': u'200',
u'timestamp': '2018-01-05T09:31:39.201000',
u'upstream_response_time': 0.0},
'event': 'nginx_event',
'timestamp': '2018-01-05T09:31:39.201000',
'type': 'metric'}
>>> input_line2 = '{ \
"remote_addr": "192.158.0.51","remote_user": "-","timestamp": "1515143686.415", \
"request": "POST /mpub?topic=heartbeat HTTP/1.1","status": "404","request_time": "0.000", \
"body_bytes_sent": "152","http_referer": "-","http_user_agent": "python-requests/2.18.4", \
"http_x_forwarded_for": "-","upstream_response_time": "-" \
}'
>>> output_line2 = nginx_access(input_line2)
>>> pprint.pprint(output_line2)
{'data': {u'body_bytes_sent': 152.0,
u'http_referer': u'-',
u'http_user_agent': u'python-requests/2.18.4',
u'http_x_forwarded_for': u'-',
u'remote_addr': u'192.158.0.51',
u'remote_user': u'-',
u'request': u'POST /mpub?topic=heartbeat HTTP/1.1',
u'request_time': 0.0,
u'status': u'404',
u'timestamp': '2018-01-05T09:14:46.415000',
u'upstream_response_time': 0.0},
'event': 'nginx_event',
'timestamp': '2018-01-05T09:14:46.415000',
'type': 'metric'}
'''
#TODO Handle nginx error logs
log = json.loads(line)
timestamp_iso = datetime.datetime.utcfromtimestamp(float(log['timestamp'])).isoformat()
log.update({'timestamp':timestamp_iso})
if '-' in log.get('upstream_response_time'):
log['upstream_response_time'] = 0.0
log['body_bytes_sent'] = float(log['body_bytes_sent'])
log['request_time'] = float(log['request_time'])
log['upstream_response_time'] = float(log['upstream_response_time'])
return dict(
timestamp=log.get('timestamp',' '),
data=log,
type='metric',
event='nginx_event',
)
|
python
|
def nginx_access(line):
'''
>>> import pprint
>>> input_line1 = '{ \
"remote_addr": "127.0.0.1","remote_user": "-","timestamp": "1515144699.201", \
"request": "GET / HTTP/1.1","status": "200","request_time": "0.000", \
"body_bytes_sent": "396","http_referer": "-","http_user_agent": "python-requests/2.18.4", \
"http_x_forwarded_for": "-","upstream_response_time": "-" \
}'
>>> output_line1 = nginx_access(input_line1)
>>> pprint.pprint(output_line1)
{'data': {u'body_bytes_sent': 396.0,
u'http_referer': u'-',
u'http_user_agent': u'python-requests/2.18.4',
u'http_x_forwarded_for': u'-',
u'remote_addr': u'127.0.0.1',
u'remote_user': u'-',
u'request': u'GET / HTTP/1.1',
u'request_time': 0.0,
u'status': u'200',
u'timestamp': '2018-01-05T09:31:39.201000',
u'upstream_response_time': 0.0},
'event': 'nginx_event',
'timestamp': '2018-01-05T09:31:39.201000',
'type': 'metric'}
>>> input_line2 = '{ \
"remote_addr": "192.158.0.51","remote_user": "-","timestamp": "1515143686.415", \
"request": "POST /mpub?topic=heartbeat HTTP/1.1","status": "404","request_time": "0.000", \
"body_bytes_sent": "152","http_referer": "-","http_user_agent": "python-requests/2.18.4", \
"http_x_forwarded_for": "-","upstream_response_time": "-" \
}'
>>> output_line2 = nginx_access(input_line2)
>>> pprint.pprint(output_line2)
{'data': {u'body_bytes_sent': 152.0,
u'http_referer': u'-',
u'http_user_agent': u'python-requests/2.18.4',
u'http_x_forwarded_for': u'-',
u'remote_addr': u'192.158.0.51',
u'remote_user': u'-',
u'request': u'POST /mpub?topic=heartbeat HTTP/1.1',
u'request_time': 0.0,
u'status': u'404',
u'timestamp': '2018-01-05T09:14:46.415000',
u'upstream_response_time': 0.0},
'event': 'nginx_event',
'timestamp': '2018-01-05T09:14:46.415000',
'type': 'metric'}
'''
#TODO Handle nginx error logs
log = json.loads(line)
timestamp_iso = datetime.datetime.utcfromtimestamp(float(log['timestamp'])).isoformat()
log.update({'timestamp':timestamp_iso})
if '-' in log.get('upstream_response_time'):
log['upstream_response_time'] = 0.0
log['body_bytes_sent'] = float(log['body_bytes_sent'])
log['request_time'] = float(log['request_time'])
log['upstream_response_time'] = float(log['upstream_response_time'])
return dict(
timestamp=log.get('timestamp',' '),
data=log,
type='metric',
event='nginx_event',
)
|
[
"def",
"nginx_access",
"(",
"line",
")",
":",
"#TODO Handle nginx error logs",
"log",
"=",
"json",
".",
"loads",
"(",
"line",
")",
"timestamp_iso",
"=",
"datetime",
".",
"datetime",
".",
"utcfromtimestamp",
"(",
"float",
"(",
"log",
"[",
"'timestamp'",
"]",
")",
")",
".",
"isoformat",
"(",
")",
"log",
".",
"update",
"(",
"{",
"'timestamp'",
":",
"timestamp_iso",
"}",
")",
"if",
"'-'",
"in",
"log",
".",
"get",
"(",
"'upstream_response_time'",
")",
":",
"log",
"[",
"'upstream_response_time'",
"]",
"=",
"0.0",
"log",
"[",
"'body_bytes_sent'",
"]",
"=",
"float",
"(",
"log",
"[",
"'body_bytes_sent'",
"]",
")",
"log",
"[",
"'request_time'",
"]",
"=",
"float",
"(",
"log",
"[",
"'request_time'",
"]",
")",
"log",
"[",
"'upstream_response_time'",
"]",
"=",
"float",
"(",
"log",
"[",
"'upstream_response_time'",
"]",
")",
"return",
"dict",
"(",
"timestamp",
"=",
"log",
".",
"get",
"(",
"'timestamp'",
",",
"' '",
")",
",",
"data",
"=",
"log",
",",
"type",
"=",
"'metric'",
",",
"event",
"=",
"'nginx_event'",
",",
")"
] |
>>> import pprint
>>> input_line1 = '{ \
"remote_addr": "127.0.0.1","remote_user": "-","timestamp": "1515144699.201", \
"request": "GET / HTTP/1.1","status": "200","request_time": "0.000", \
"body_bytes_sent": "396","http_referer": "-","http_user_agent": "python-requests/2.18.4", \
"http_x_forwarded_for": "-","upstream_response_time": "-" \
}'
>>> output_line1 = nginx_access(input_line1)
>>> pprint.pprint(output_line1)
{'data': {u'body_bytes_sent': 396.0,
u'http_referer': u'-',
u'http_user_agent': u'python-requests/2.18.4',
u'http_x_forwarded_for': u'-',
u'remote_addr': u'127.0.0.1',
u'remote_user': u'-',
u'request': u'GET / HTTP/1.1',
u'request_time': 0.0,
u'status': u'200',
u'timestamp': '2018-01-05T09:31:39.201000',
u'upstream_response_time': 0.0},
'event': 'nginx_event',
'timestamp': '2018-01-05T09:31:39.201000',
'type': 'metric'}
>>> input_line2 = '{ \
"remote_addr": "192.158.0.51","remote_user": "-","timestamp": "1515143686.415", \
"request": "POST /mpub?topic=heartbeat HTTP/1.1","status": "404","request_time": "0.000", \
"body_bytes_sent": "152","http_referer": "-","http_user_agent": "python-requests/2.18.4", \
"http_x_forwarded_for": "-","upstream_response_time": "-" \
}'
>>> output_line2 = nginx_access(input_line2)
>>> pprint.pprint(output_line2)
{'data': {u'body_bytes_sent': 152.0,
u'http_referer': u'-',
u'http_user_agent': u'python-requests/2.18.4',
u'http_x_forwarded_for': u'-',
u'remote_addr': u'192.158.0.51',
u'remote_user': u'-',
u'request': u'POST /mpub?topic=heartbeat HTTP/1.1',
u'request_time': 0.0,
u'status': u'404',
u'timestamp': '2018-01-05T09:14:46.415000',
u'upstream_response_time': 0.0},
'event': 'nginx_event',
'timestamp': '2018-01-05T09:14:46.415000',
'type': 'metric'}
|
[
">>>",
"import",
"pprint",
">>>",
"input_line1",
"=",
"{",
"\\",
"remote_addr",
":",
"127",
".",
"0",
".",
"0",
".",
"1",
"remote_user",
":",
"-",
"timestamp",
":",
"1515144699",
".",
"201",
"\\",
"request",
":",
"GET",
"/",
"HTTP",
"/",
"1",
".",
"1",
"status",
":",
"200",
"request_time",
":",
"0",
".",
"000",
"\\",
"body_bytes_sent",
":",
"396",
"http_referer",
":",
"-",
"http_user_agent",
":",
"python",
"-",
"requests",
"/",
"2",
".",
"18",
".",
"4",
"\\",
"http_x_forwarded_for",
":",
"-",
"upstream_response_time",
":",
"-",
"\\",
"}",
">>>",
"output_line1",
"=",
"nginx_access",
"(",
"input_line1",
")",
">>>",
"pprint",
".",
"pprint",
"(",
"output_line1",
")",
"{",
"data",
":",
"{",
"u",
"body_bytes_sent",
":",
"396",
".",
"0",
"u",
"http_referer",
":",
"u",
"-",
"u",
"http_user_agent",
":",
"u",
"python",
"-",
"requests",
"/",
"2",
".",
"18",
".",
"4",
"u",
"http_x_forwarded_for",
":",
"u",
"-",
"u",
"remote_addr",
":",
"u",
"127",
".",
"0",
".",
"0",
".",
"1",
"u",
"remote_user",
":",
"u",
"-",
"u",
"request",
":",
"u",
"GET",
"/",
"HTTP",
"/",
"1",
".",
"1",
"u",
"request_time",
":",
"0",
".",
"0",
"u",
"status",
":",
"u",
"200",
"u",
"timestamp",
":",
"2018",
"-",
"01",
"-",
"05T09",
":",
"31",
":",
"39",
".",
"201000",
"u",
"upstream_response_time",
":",
"0",
".",
"0",
"}",
"event",
":",
"nginx_event",
"timestamp",
":",
"2018",
"-",
"01",
"-",
"05T09",
":",
"31",
":",
"39",
".",
"201000",
"type",
":",
"metric",
"}"
] |
train
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L96-L160
|
deep-compute/logagg
|
logagg/formatters.py
|
mongodb
|
def mongodb(line):
'''
>>> import pprint
>>> input_line1 = '2017-08-17T07:56:33.489+0200 I REPL [signalProcessingThread] shutting down replication subsystems'
>>> output_line1 = mongodb(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'component': 'REPL',
'context': '[signalProcessingThread]',
'message': 'shutting down replication subsystems',
'severity': 'I',
'timestamp': '2017-08-17T07:56:33.489+0200'},
'timestamp': '2017-08-17T07:56:33.489+0200',
'type': 'log'}
>>> input_line2 = '2017-08-17T07:56:33.515+0200 W NETWORK [initandlisten] No primary detected for set confsvr_repl1'
>>> output_line2 = mongodb(input_line2)
>>> pprint.pprint(output_line2)
{'data': {'component': 'NETWORK',
'context': '[initandlisten]',
'message': 'No primary detected for set confsvr_repl1',
'severity': 'W',
'timestamp': '2017-08-17T07:56:33.515+0200'},
'timestamp': '2017-08-17T07:56:33.515+0200',
'type': 'log'}
'''
keys = ['timestamp', 'severity', 'component', 'context', 'message']
values = re.split(r'\s+', line, maxsplit=4)
mongodb_log = dict(zip(keys,values))
return dict(
timestamp=values[0],
data=mongodb_log,
type='log',
)
|
python
|
def mongodb(line):
'''
>>> import pprint
>>> input_line1 = '2017-08-17T07:56:33.489+0200 I REPL [signalProcessingThread] shutting down replication subsystems'
>>> output_line1 = mongodb(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'component': 'REPL',
'context': '[signalProcessingThread]',
'message': 'shutting down replication subsystems',
'severity': 'I',
'timestamp': '2017-08-17T07:56:33.489+0200'},
'timestamp': '2017-08-17T07:56:33.489+0200',
'type': 'log'}
>>> input_line2 = '2017-08-17T07:56:33.515+0200 W NETWORK [initandlisten] No primary detected for set confsvr_repl1'
>>> output_line2 = mongodb(input_line2)
>>> pprint.pprint(output_line2)
{'data': {'component': 'NETWORK',
'context': '[initandlisten]',
'message': 'No primary detected for set confsvr_repl1',
'severity': 'W',
'timestamp': '2017-08-17T07:56:33.515+0200'},
'timestamp': '2017-08-17T07:56:33.515+0200',
'type': 'log'}
'''
keys = ['timestamp', 'severity', 'component', 'context', 'message']
values = re.split(r'\s+', line, maxsplit=4)
mongodb_log = dict(zip(keys,values))
return dict(
timestamp=values[0],
data=mongodb_log,
type='log',
)
|
[
"def",
"mongodb",
"(",
"line",
")",
":",
"keys",
"=",
"[",
"'timestamp'",
",",
"'severity'",
",",
"'component'",
",",
"'context'",
",",
"'message'",
"]",
"values",
"=",
"re",
".",
"split",
"(",
"r'\\s+'",
",",
"line",
",",
"maxsplit",
"=",
"4",
")",
"mongodb_log",
"=",
"dict",
"(",
"zip",
"(",
"keys",
",",
"values",
")",
")",
"return",
"dict",
"(",
"timestamp",
"=",
"values",
"[",
"0",
"]",
",",
"data",
"=",
"mongodb_log",
",",
"type",
"=",
"'log'",
",",
")"
] |
>>> import pprint
>>> input_line1 = '2017-08-17T07:56:33.489+0200 I REPL [signalProcessingThread] shutting down replication subsystems'
>>> output_line1 = mongodb(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'component': 'REPL',
'context': '[signalProcessingThread]',
'message': 'shutting down replication subsystems',
'severity': 'I',
'timestamp': '2017-08-17T07:56:33.489+0200'},
'timestamp': '2017-08-17T07:56:33.489+0200',
'type': 'log'}
>>> input_line2 = '2017-08-17T07:56:33.515+0200 W NETWORK [initandlisten] No primary detected for set confsvr_repl1'
>>> output_line2 = mongodb(input_line2)
>>> pprint.pprint(output_line2)
{'data': {'component': 'NETWORK',
'context': '[initandlisten]',
'message': 'No primary detected for set confsvr_repl1',
'severity': 'W',
'timestamp': '2017-08-17T07:56:33.515+0200'},
'timestamp': '2017-08-17T07:56:33.515+0200',
'type': 'log'}
|
[
">>>",
"import",
"pprint",
">>>",
"input_line1",
"=",
"2017",
"-",
"08",
"-",
"17T07",
":",
"56",
":",
"33",
".",
"489",
"+",
"0200",
"I",
"REPL",
"[",
"signalProcessingThread",
"]",
"shutting",
"down",
"replication",
"subsystems",
">>>",
"output_line1",
"=",
"mongodb",
"(",
"input_line1",
")",
">>>",
"pprint",
".",
"pprint",
"(",
"output_line1",
")",
"{",
"data",
":",
"{",
"component",
":",
"REPL",
"context",
":",
"[",
"signalProcessingThread",
"]",
"message",
":",
"shutting",
"down",
"replication",
"subsystems",
"severity",
":",
"I",
"timestamp",
":",
"2017",
"-",
"08",
"-",
"17T07",
":",
"56",
":",
"33",
".",
"489",
"+",
"0200",
"}",
"timestamp",
":",
"2017",
"-",
"08",
"-",
"17T07",
":",
"56",
":",
"33",
".",
"489",
"+",
"0200",
"type",
":",
"log",
"}"
] |
train
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L162-L196
|
deep-compute/logagg
|
logagg/formatters.py
|
django
|
def django(line):
'''
>>> import pprint
>>> input_line1 = '[23/Aug/2017 11:35:25] INFO [app.middleware_log_req:50]View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }'
>>> output_line1 = django(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'loglevel': 'INFO',
'logname': '[app.middleware_log_req:50]',
'message': 'View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }',
'timestamp': '2017-08-23T11:35:25'},
'level': 'INFO',
'timestamp': '2017-08-23T11:35:25'}
>>> input_line2 = '[22/Sep/2017 06:32:15] INFO [app.function:6022] {"UUID": "c47f3530-9f5f-11e7-a559-917d011459f7", "timestamp":1506061932546, "misc": {"status": 200, "ready_state": 4, "end_time_ms": 1506061932546, "url": "/api/function?", "start_time_ms": 1506061932113, "response_length": 31, "status_message": "OK", "request_time_ms": 433}, "user": "root", "host_url": "localhost:8888", "message": "ajax success"}'
>>> output_line2 = django(input_line2)
>>> pprint.pprint(output_line2)
{'data': {'loglevel': 'INFO',
'logname': '[app.function:6022]',
'message': {u'UUID': u'c47f3530-9f5f-11e7-a559-917d011459f7',
u'host_url': u'localhost:8888',
u'message': u'ajax success',
u'misc': {u'end_time_ms': 1506061932546L,
u'ready_state': 4,
u'request_time_ms': 433,
u'response_length': 31,
u'start_time_ms': 1506061932113L,
u'status': 200,
u'status_message': u'OK',
u'url': u'/api/function?'},
u'timestamp': 1506061932546L,
u'user': u'root'},
'timestamp': '2017-09-22T06:32:15'},
'level': 'INFO',
'timestamp': '2017-09-22T06:32:15'}
Case2:
[18/Sep/2017 05:40:36] ERROR [app.apps:78] failed to get the record, collection = Collection(Database(MongoClient(host=['localhost:27017'], document_class=dict, tz_aware=False, connect=True, serverselectiontimeoutms=3000), u'collection_cache'), u'function_dummy_version')
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/mongo_cache/mongocache.py", line 70, in __getitem__
result = self.collection.find_one({"_id": key})
OperationFailure: not authorized on collection_cache to execute command { find: "function", filter: { _id: "zydelig-cosine-20" }, limit: 1, singleBatch: true }
'''
#TODO we need to handle case2 logs
data = {}
log = re.findall(r'^(\[\d+/\w+/\d+ \d+:\d+:\d+\].*)', line)
if len(log) == 1:
data['timestamp'] = datetime.datetime.strptime(re.findall(r'(\d+/\w+/\d+ \d+:\d+:\d+)',\
log[0])[0],"%d/%b/%Y %H:%M:%S").isoformat()
data['loglevel'] = re.findall('[A-Z]+', log[0])[1]
data['logname'] = re.findall('\[\D+.\w+:\d+\]', log[0])[0]
message = re.findall('\{.+\}', log[0])
try:
if len(message) > 0:
message = json.loads(message[0])
else:
message = re.split(']', log[0])
message = ''.join(message[2:])
except ValueError:
message = re.split(']', log[0])
message = ''.join(message[2:])
data['message'] = message
return dict(
timestamp=data['timestamp'],
level=data['loglevel'],
data=data,
)
else:
return dict(
timestamp=datetime.datetime.isoformat(datetime.datetime.utcnow()),
data={raw:line}
)
|
python
|
def django(line):
'''
>>> import pprint
>>> input_line1 = '[23/Aug/2017 11:35:25] INFO [app.middleware_log_req:50]View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }'
>>> output_line1 = django(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'loglevel': 'INFO',
'logname': '[app.middleware_log_req:50]',
'message': 'View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }',
'timestamp': '2017-08-23T11:35:25'},
'level': 'INFO',
'timestamp': '2017-08-23T11:35:25'}
>>> input_line2 = '[22/Sep/2017 06:32:15] INFO [app.function:6022] {"UUID": "c47f3530-9f5f-11e7-a559-917d011459f7", "timestamp":1506061932546, "misc": {"status": 200, "ready_state": 4, "end_time_ms": 1506061932546, "url": "/api/function?", "start_time_ms": 1506061932113, "response_length": 31, "status_message": "OK", "request_time_ms": 433}, "user": "root", "host_url": "localhost:8888", "message": "ajax success"}'
>>> output_line2 = django(input_line2)
>>> pprint.pprint(output_line2)
{'data': {'loglevel': 'INFO',
'logname': '[app.function:6022]',
'message': {u'UUID': u'c47f3530-9f5f-11e7-a559-917d011459f7',
u'host_url': u'localhost:8888',
u'message': u'ajax success',
u'misc': {u'end_time_ms': 1506061932546L,
u'ready_state': 4,
u'request_time_ms': 433,
u'response_length': 31,
u'start_time_ms': 1506061932113L,
u'status': 200,
u'status_message': u'OK',
u'url': u'/api/function?'},
u'timestamp': 1506061932546L,
u'user': u'root'},
'timestamp': '2017-09-22T06:32:15'},
'level': 'INFO',
'timestamp': '2017-09-22T06:32:15'}
Case2:
[18/Sep/2017 05:40:36] ERROR [app.apps:78] failed to get the record, collection = Collection(Database(MongoClient(host=['localhost:27017'], document_class=dict, tz_aware=False, connect=True, serverselectiontimeoutms=3000), u'collection_cache'), u'function_dummy_version')
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/mongo_cache/mongocache.py", line 70, in __getitem__
result = self.collection.find_one({"_id": key})
OperationFailure: not authorized on collection_cache to execute command { find: "function", filter: { _id: "zydelig-cosine-20" }, limit: 1, singleBatch: true }
'''
#TODO we need to handle case2 logs
data = {}
log = re.findall(r'^(\[\d+/\w+/\d+ \d+:\d+:\d+\].*)', line)
if len(log) == 1:
data['timestamp'] = datetime.datetime.strptime(re.findall(r'(\d+/\w+/\d+ \d+:\d+:\d+)',\
log[0])[0],"%d/%b/%Y %H:%M:%S").isoformat()
data['loglevel'] = re.findall('[A-Z]+', log[0])[1]
data['logname'] = re.findall('\[\D+.\w+:\d+\]', log[0])[0]
message = re.findall('\{.+\}', log[0])
try:
if len(message) > 0:
message = json.loads(message[0])
else:
message = re.split(']', log[0])
message = ''.join(message[2:])
except ValueError:
message = re.split(']', log[0])
message = ''.join(message[2:])
data['message'] = message
return dict(
timestamp=data['timestamp'],
level=data['loglevel'],
data=data,
)
else:
return dict(
timestamp=datetime.datetime.isoformat(datetime.datetime.utcnow()),
data={raw:line}
)
|
[
"def",
"django",
"(",
"line",
")",
":",
"#TODO we need to handle case2 logs",
"data",
"=",
"{",
"}",
"log",
"=",
"re",
".",
"findall",
"(",
"r'^(\\[\\d+/\\w+/\\d+ \\d+:\\d+:\\d+\\].*)'",
",",
"line",
")",
"if",
"len",
"(",
"log",
")",
"==",
"1",
":",
"data",
"[",
"'timestamp'",
"]",
"=",
"datetime",
".",
"datetime",
".",
"strptime",
"(",
"re",
".",
"findall",
"(",
"r'(\\d+/\\w+/\\d+ \\d+:\\d+:\\d+)'",
",",
"log",
"[",
"0",
"]",
")",
"[",
"0",
"]",
",",
"\"%d/%b/%Y %H:%M:%S\"",
")",
".",
"isoformat",
"(",
")",
"data",
"[",
"'loglevel'",
"]",
"=",
"re",
".",
"findall",
"(",
"'[A-Z]+'",
",",
"log",
"[",
"0",
"]",
")",
"[",
"1",
"]",
"data",
"[",
"'logname'",
"]",
"=",
"re",
".",
"findall",
"(",
"'\\[\\D+.\\w+:\\d+\\]'",
",",
"log",
"[",
"0",
"]",
")",
"[",
"0",
"]",
"message",
"=",
"re",
".",
"findall",
"(",
"'\\{.+\\}'",
",",
"log",
"[",
"0",
"]",
")",
"try",
":",
"if",
"len",
"(",
"message",
")",
">",
"0",
":",
"message",
"=",
"json",
".",
"loads",
"(",
"message",
"[",
"0",
"]",
")",
"else",
":",
"message",
"=",
"re",
".",
"split",
"(",
"']'",
",",
"log",
"[",
"0",
"]",
")",
"message",
"=",
"''",
".",
"join",
"(",
"message",
"[",
"2",
":",
"]",
")",
"except",
"ValueError",
":",
"message",
"=",
"re",
".",
"split",
"(",
"']'",
",",
"log",
"[",
"0",
"]",
")",
"message",
"=",
"''",
".",
"join",
"(",
"message",
"[",
"2",
":",
"]",
")",
"data",
"[",
"'message'",
"]",
"=",
"message",
"return",
"dict",
"(",
"timestamp",
"=",
"data",
"[",
"'timestamp'",
"]",
",",
"level",
"=",
"data",
"[",
"'loglevel'",
"]",
",",
"data",
"=",
"data",
",",
")",
"else",
":",
"return",
"dict",
"(",
"timestamp",
"=",
"datetime",
".",
"datetime",
".",
"isoformat",
"(",
"datetime",
".",
"datetime",
".",
"utcnow",
"(",
")",
")",
",",
"data",
"=",
"{",
"raw",
":",
"line",
"}",
")"
] |
>>> import pprint
>>> input_line1 = '[23/Aug/2017 11:35:25] INFO [app.middleware_log_req:50]View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }'
>>> output_line1 = django(input_line1)
>>> pprint.pprint(output_line1)
{'data': {'loglevel': 'INFO',
'logname': '[app.middleware_log_req:50]',
'message': 'View func called:{"exception": null,"processing_time": 0.00011801719665527344, "url": "<url>",host": "localhost", "user": "testing", "post_contents": "", "method": "POST" }',
'timestamp': '2017-08-23T11:35:25'},
'level': 'INFO',
'timestamp': '2017-08-23T11:35:25'}
>>> input_line2 = '[22/Sep/2017 06:32:15] INFO [app.function:6022] {"UUID": "c47f3530-9f5f-11e7-a559-917d011459f7", "timestamp":1506061932546, "misc": {"status": 200, "ready_state": 4, "end_time_ms": 1506061932546, "url": "/api/function?", "start_time_ms": 1506061932113, "response_length": 31, "status_message": "OK", "request_time_ms": 433}, "user": "root", "host_url": "localhost:8888", "message": "ajax success"}'
>>> output_line2 = django(input_line2)
>>> pprint.pprint(output_line2)
{'data': {'loglevel': 'INFO',
'logname': '[app.function:6022]',
'message': {u'UUID': u'c47f3530-9f5f-11e7-a559-917d011459f7',
u'host_url': u'localhost:8888',
u'message': u'ajax success',
u'misc': {u'end_time_ms': 1506061932546L,
u'ready_state': 4,
u'request_time_ms': 433,
u'response_length': 31,
u'start_time_ms': 1506061932113L,
u'status': 200,
u'status_message': u'OK',
u'url': u'/api/function?'},
u'timestamp': 1506061932546L,
u'user': u'root'},
'timestamp': '2017-09-22T06:32:15'},
'level': 'INFO',
'timestamp': '2017-09-22T06:32:15'}
Case2:
[18/Sep/2017 05:40:36] ERROR [app.apps:78] failed to get the record, collection = Collection(Database(MongoClient(host=['localhost:27017'], document_class=dict, tz_aware=False, connect=True, serverselectiontimeoutms=3000), u'collection_cache'), u'function_dummy_version')
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/mongo_cache/mongocache.py", line 70, in __getitem__
result = self.collection.find_one({"_id": key})
OperationFailure: not authorized on collection_cache to execute command { find: "function", filter: { _id: "zydelig-cosine-20" }, limit: 1, singleBatch: true }
|
[
">>>",
"import",
"pprint",
">>>",
"input_line1",
"=",
"[",
"23",
"/",
"Aug",
"/",
"2017",
"11",
":",
"35",
":",
"25",
"]",
"INFO",
"[",
"app",
".",
"middleware_log_req",
":",
"50",
"]",
"View",
"func",
"called",
":",
"{",
"exception",
":",
"null",
"processing_time",
":",
"0",
".",
"00011801719665527344",
"url",
":",
"<url",
">",
"host",
":",
"localhost",
"user",
":",
"testing",
"post_contents",
":",
"method",
":",
"POST",
"}",
">>>",
"output_line1",
"=",
"django",
"(",
"input_line1",
")",
">>>",
"pprint",
".",
"pprint",
"(",
"output_line1",
")",
"{",
"data",
":",
"{",
"loglevel",
":",
"INFO",
"logname",
":",
"[",
"app",
".",
"middleware_log_req",
":",
"50",
"]",
"message",
":",
"View",
"func",
"called",
":",
"{",
"exception",
":",
"null",
"processing_time",
":",
"0",
".",
"00011801719665527344",
"url",
":",
"<url",
">",
"host",
":",
"localhost",
"user",
":",
"testing",
"post_contents",
":",
"method",
":",
"POST",
"}",
"timestamp",
":",
"2017",
"-",
"08",
"-",
"23T11",
":",
"35",
":",
"25",
"}",
"level",
":",
"INFO",
"timestamp",
":",
"2017",
"-",
"08",
"-",
"23T11",
":",
"35",
":",
"25",
"}"
] |
train
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L199-L271
|
deep-compute/logagg
|
logagg/formatters.py
|
basescript
|
def basescript(line):
'''
>>> import pprint
>>> input_line = '{"level": "warning", "timestamp": "2018-02-07T06:37:00.297610Z", "event": "exited via keyboard interrupt", "type": "log", "id": "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_": {"ln": 58, "file": "/usr/local/lib/python2.7/dist-packages/basescript/basescript.py", "name": "basescript.basescript", "fn": "start"}}'
>>> output_line1 = basescript(input_line)
>>> pprint.pprint(output_line1)
{'data': {u'_': {u'file': u'/usr/local/lib/python2.7/dist-packages/basescript/basescript.py',
u'fn': u'start',
u'ln': 58,
u'name': u'basescript.basescript'},
u'event': u'exited via keyboard interrupt',
u'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65',
u'level': u'warning',
u'timestamp': u'2018-02-07T06:37:00.297610Z',
u'type': u'log'},
'event': u'exited via keyboard interrupt',
'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65',
'level': u'warning',
'timestamp': u'2018-02-07T06:37:00.297610Z',
'type': u'log'}
'''
log = json.loads(line)
return dict(
timestamp=log['timestamp'],
data=log,
id=log['id'],
type=log['type'],
level=log['level'],
event=log['event']
)
|
python
|
def basescript(line):
'''
>>> import pprint
>>> input_line = '{"level": "warning", "timestamp": "2018-02-07T06:37:00.297610Z", "event": "exited via keyboard interrupt", "type": "log", "id": "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_": {"ln": 58, "file": "/usr/local/lib/python2.7/dist-packages/basescript/basescript.py", "name": "basescript.basescript", "fn": "start"}}'
>>> output_line1 = basescript(input_line)
>>> pprint.pprint(output_line1)
{'data': {u'_': {u'file': u'/usr/local/lib/python2.7/dist-packages/basescript/basescript.py',
u'fn': u'start',
u'ln': 58,
u'name': u'basescript.basescript'},
u'event': u'exited via keyboard interrupt',
u'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65',
u'level': u'warning',
u'timestamp': u'2018-02-07T06:37:00.297610Z',
u'type': u'log'},
'event': u'exited via keyboard interrupt',
'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65',
'level': u'warning',
'timestamp': u'2018-02-07T06:37:00.297610Z',
'type': u'log'}
'''
log = json.loads(line)
return dict(
timestamp=log['timestamp'],
data=log,
id=log['id'],
type=log['type'],
level=log['level'],
event=log['event']
)
|
[
"def",
"basescript",
"(",
"line",
")",
":",
"log",
"=",
"json",
".",
"loads",
"(",
"line",
")",
"return",
"dict",
"(",
"timestamp",
"=",
"log",
"[",
"'timestamp'",
"]",
",",
"data",
"=",
"log",
",",
"id",
"=",
"log",
"[",
"'id'",
"]",
",",
"type",
"=",
"log",
"[",
"'type'",
"]",
",",
"level",
"=",
"log",
"[",
"'level'",
"]",
",",
"event",
"=",
"log",
"[",
"'event'",
"]",
")"
] |
>>> import pprint
>>> input_line = '{"level": "warning", "timestamp": "2018-02-07T06:37:00.297610Z", "event": "exited via keyboard interrupt", "type": "log", "id": "20180207T063700_4d03fe800bd111e89ecb96000007bc65", "_": {"ln": 58, "file": "/usr/local/lib/python2.7/dist-packages/basescript/basescript.py", "name": "basescript.basescript", "fn": "start"}}'
>>> output_line1 = basescript(input_line)
>>> pprint.pprint(output_line1)
{'data': {u'_': {u'file': u'/usr/local/lib/python2.7/dist-packages/basescript/basescript.py',
u'fn': u'start',
u'ln': 58,
u'name': u'basescript.basescript'},
u'event': u'exited via keyboard interrupt',
u'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65',
u'level': u'warning',
u'timestamp': u'2018-02-07T06:37:00.297610Z',
u'type': u'log'},
'event': u'exited via keyboard interrupt',
'id': u'20180207T063700_4d03fe800bd111e89ecb96000007bc65',
'level': u'warning',
'timestamp': u'2018-02-07T06:37:00.297610Z',
'type': u'log'}
|
[
">>>",
"import",
"pprint",
">>>",
"input_line",
"=",
"{",
"level",
":",
"warning",
"timestamp",
":",
"2018",
"-",
"02",
"-",
"07T06",
":",
"37",
":",
"00",
".",
"297610Z",
"event",
":",
"exited",
"via",
"keyboard",
"interrupt",
"type",
":",
"log",
"id",
":",
"20180207T063700_4d03fe800bd111e89ecb96000007bc65",
"_",
":",
"{",
"ln",
":",
"58",
"file",
":",
"/",
"usr",
"/",
"local",
"/",
"lib",
"/",
"python2",
".",
"7",
"/",
"dist",
"-",
"packages",
"/",
"basescript",
"/",
"basescript",
".",
"py",
"name",
":",
"basescript",
".",
"basescript",
"fn",
":",
"start",
"}}",
">>>",
"output_line1",
"=",
"basescript",
"(",
"input_line",
")",
">>>",
"pprint",
".",
"pprint",
"(",
"output_line1",
")",
"{",
"data",
":",
"{",
"u",
"_",
":",
"{",
"u",
"file",
":",
"u",
"/",
"usr",
"/",
"local",
"/",
"lib",
"/",
"python2",
".",
"7",
"/",
"dist",
"-",
"packages",
"/",
"basescript",
"/",
"basescript",
".",
"py",
"u",
"fn",
":",
"u",
"start",
"u",
"ln",
":",
"58",
"u",
"name",
":",
"u",
"basescript",
".",
"basescript",
"}",
"u",
"event",
":",
"u",
"exited",
"via",
"keyboard",
"interrupt",
"u",
"id",
":",
"u",
"20180207T063700_4d03fe800bd111e89ecb96000007bc65",
"u",
"level",
":",
"u",
"warning",
"u",
"timestamp",
":",
"u",
"2018",
"-",
"02",
"-",
"07T06",
":",
"37",
":",
"00",
".",
"297610Z",
"u",
"type",
":",
"u",
"log",
"}",
"event",
":",
"u",
"exited",
"via",
"keyboard",
"interrupt",
"id",
":",
"u",
"20180207T063700_4d03fe800bd111e89ecb96000007bc65",
"level",
":",
"u",
"warning",
"timestamp",
":",
"u",
"2018",
"-",
"02",
"-",
"07T06",
":",
"37",
":",
"00",
".",
"297610Z",
"type",
":",
"u",
"log",
"}"
] |
train
|
https://github.com/deep-compute/logagg/blob/7863bc1b5ddf3e67c4d4b55746799304180589a0/logagg/formatters.py#L273-L304
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.