Unnamed: 0 int64 0 10k | repository_name stringlengths 7 54 | func_path_in_repository stringlengths 5 223 | func_name stringlengths 1 134 | whole_func_string stringlengths 100 30.3k | language stringclasses 1
value | func_code_string stringlengths 100 30.3k | func_code_tokens stringlengths 138 33.2k | func_documentation_string stringlengths 1 15k | func_documentation_tokens stringlengths 5 5.14k | split_name stringclasses 1
value | func_code_url stringlengths 91 315 |
|---|---|---|---|---|---|---|---|---|---|---|---|
8,200 | inveniosoftware/invenio-communities | invenio_communities/views/ui.py | curate | def curate(community):
"""Index page with uploader and list of existing depositions.
:param community_id: ID of the community to curate.
"""
if request.method == 'POST':
action = request.json.get('action')
recid = request.json.get('recid')
# 'recid' is mandatory
if not ... | python | def curate(community):
"""Index page with uploader and list of existing depositions.
:param community_id: ID of the community to curate.
"""
if request.method == 'POST':
action = request.json.get('action')
recid = request.json.get('recid')
# 'recid' is mandatory
if not ... | ['def', 'curate', '(', 'community', ')', ':', 'if', 'request', '.', 'method', '==', "'POST'", ':', 'action', '=', 'request', '.', 'json', '.', 'get', '(', "'action'", ')', 'recid', '=', 'request', '.', 'json', '.', 'get', '(', "'recid'", ')', "# 'recid' is mandatory", 'if', 'not', 'recid', ':', 'abort', '(', '400', ')'... | Index page with uploader and list of existing depositions.
:param community_id: ID of the community to curate. | ['Index', 'page', 'with', 'uploader', 'and', 'list', 'of', 'existing', 'depositions', '.'] | train | https://github.com/inveniosoftware/invenio-communities/blob/5c4de6783724d276ae1b6dd13a399a9e22fadc7a/invenio_communities/views/ui.py#L278-L315 |
8,201 | rosenbrockc/fortpy | fortpy/elements.py | Module.completions | def completions(self, symbol, attribute, recursive = False):
"""Finds all possible symbol completions of the given symbol that belong
to this module and its dependencies.
:arg symbol: the code symbol that needs to be completed.
:arg attribute: one of ['dependencies', 'publics', 'members... | python | def completions(self, symbol, attribute, recursive = False):
"""Finds all possible symbol completions of the given symbol that belong
to this module and its dependencies.
:arg symbol: the code symbol that needs to be completed.
:arg attribute: one of ['dependencies', 'publics', 'members... | ['def', 'completions', '(', 'self', ',', 'symbol', ',', 'attribute', ',', 'recursive', '=', 'False', ')', ':', 'possible', '=', '[', ']', 'for', 'ekey', 'in', 'self', '.', 'collection', '(', 'attribute', ')', ':', 'if', 'symbol', 'in', 'ekey', ':', 'possible', '.', 'append', '(', 'ekey', ')', '#Try this out on all the ... | Finds all possible symbol completions of the given symbol that belong
to this module and its dependencies.
:arg symbol: the code symbol that needs to be completed.
:arg attribute: one of ['dependencies', 'publics', 'members',
'types', 'executables'] for specifying which collections t... | ['Finds', 'all', 'possible', 'symbol', 'completions', 'of', 'the', 'given', 'symbol', 'that', 'belong', 'to', 'this', 'module', 'and', 'its', 'dependencies', '.'] | train | https://github.com/rosenbrockc/fortpy/blob/1ed0757c52d549e41d9d44bdea68cb89529293a5/fortpy/elements.py#L1577-L1600 |
8,202 | rgs1/zk_shell | zk_shell/shell.py | Shell.do_rm | def do_rm(self, params):
"""
\x1b[1mNAME\x1b[0m
rm - Remove the znode
\x1b[1mSYNOPSIS\x1b[0m
rm <path> [path] [path] ... [path]
\x1b[1mEXAMPLES\x1b[0m
> rm /foo
> rm /foo /bar
"""
for path in params.paths:
try:
self.client_context.de... | python | def do_rm(self, params):
"""
\x1b[1mNAME\x1b[0m
rm - Remove the znode
\x1b[1mSYNOPSIS\x1b[0m
rm <path> [path] [path] ... [path]
\x1b[1mEXAMPLES\x1b[0m
> rm /foo
> rm /foo /bar
"""
for path in params.paths:
try:
self.client_context.de... | ['def', 'do_rm', '(', 'self', ',', 'params', ')', ':', 'for', 'path', 'in', 'params', '.', 'paths', ':', 'try', ':', 'self', '.', 'client_context', '.', 'delete', '(', 'path', ')', 'except', 'NotEmptyError', ':', 'self', '.', 'show_output', '(', '"%s is not empty."', ',', 'path', ')', 'except', 'NoNodeError', ':', 'sel... | \x1b[1mNAME\x1b[0m
rm - Remove the znode
\x1b[1mSYNOPSIS\x1b[0m
rm <path> [path] [path] ... [path]
\x1b[1mEXAMPLES\x1b[0m
> rm /foo
> rm /foo /bar | ['\\', 'x1b', '[', '1mNAME', '\\', 'x1b', '[', '0m', 'rm', '-', 'Remove', 'the', 'znode'] | train | https://github.com/rgs1/zk_shell/blob/bbf34fdfcf1f81100e2a5816fad8af6afc782a54/zk_shell/shell.py#L1309-L1328 |
8,203 | stevearc/dql | dql/cli.py | DQLClient.do_watch | def do_watch(self, *args):
""" Watch Dynamo tables consumed capacity """
tables = []
if not self.engine.cached_descriptions:
self.engine.describe_all()
all_tables = list(self.engine.cached_descriptions)
for arg in args:
candidates = set((t for t in all_tab... | python | def do_watch(self, *args):
""" Watch Dynamo tables consumed capacity """
tables = []
if not self.engine.cached_descriptions:
self.engine.describe_all()
all_tables = list(self.engine.cached_descriptions)
for arg in args:
candidates = set((t for t in all_tab... | ['def', 'do_watch', '(', 'self', ',', '*', 'args', ')', ':', 'tables', '=', '[', ']', 'if', 'not', 'self', '.', 'engine', '.', 'cached_descriptions', ':', 'self', '.', 'engine', '.', 'describe_all', '(', ')', 'all_tables', '=', 'list', '(', 'self', '.', 'engine', '.', 'cached_descriptions', ')', 'for', 'arg', 'in', 'ar... | Watch Dynamo tables consumed capacity | ['Watch', 'Dynamo', 'tables', 'consumed', 'capacity'] | train | https://github.com/stevearc/dql/blob/e9d3aa22873076dae5ebd02e35318aa996b1e56a/dql/cli.py#L400-L413 |
8,204 | nuagenetworks/bambou | bambou/nurest_object.py | NURESTObject.children_rest_names | def children_rest_names(self):
""" Gets the list of all possible children ReST names.
Returns:
list: list containing all possible rest names as string
Example:
>>> entity = NUEntity()
>>> entity.children_rest_names
["foo",... | python | def children_rest_names(self):
""" Gets the list of all possible children ReST names.
Returns:
list: list containing all possible rest names as string
Example:
>>> entity = NUEntity()
>>> entity.children_rest_names
["foo",... | ['def', 'children_rest_names', '(', 'self', ')', ':', 'names', '=', '[', ']', 'for', 'fetcher', 'in', 'self', '.', 'fetchers', ':', 'names', '.', 'append', '(', 'fetcher', '.', '__class__', '.', 'managed_object_rest_name', '(', ')', ')', 'return', 'names'] | Gets the list of all possible children ReST names.
Returns:
list: list containing all possible rest names as string
Example:
>>> entity = NUEntity()
>>> entity.children_rest_names
["foo", "bar"] | ['Gets', 'the', 'list', 'of', 'all', 'possible', 'children', 'ReST', 'names', '.'] | train | https://github.com/nuagenetworks/bambou/blob/d334fea23e384d3df8e552fe1849ad707941c666/bambou/nurest_object.py#L266-L283 |
8,205 | santosjorge/cufflinks | cufflinks/quant_figure.py | QuantFig.update | def update(self,**kwargs):
"""
Updates the values for a QuantFigure
The key-values are automatically assigned to the correct
section of the QuantFigure
"""
if 'columns' in kwargs:
self._d=ta._ohlc_dict(self.df,columns=kwargs.pop('columns',None))
schema=self._get_schema()
annotations=kwargs.pop('anno... | python | def update(self,**kwargs):
"""
Updates the values for a QuantFigure
The key-values are automatically assigned to the correct
section of the QuantFigure
"""
if 'columns' in kwargs:
self._d=ta._ohlc_dict(self.df,columns=kwargs.pop('columns',None))
schema=self._get_schema()
annotations=kwargs.pop('anno... | ['def', 'update', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'if', "'columns'", 'in', 'kwargs', ':', 'self', '.', '_d', '=', 'ta', '.', '_ohlc_dict', '(', 'self', '.', 'df', ',', 'columns', '=', 'kwargs', '.', 'pop', '(', "'columns'", ',', 'None', ')', ')', 'schema', '=', 'self', '.', '_get_schema', '(', ')', 'ann... | Updates the values for a QuantFigure
The key-values are automatically assigned to the correct
section of the QuantFigure | ['Updates', 'the', 'values', 'for', 'a', 'QuantFigure', 'The', 'key', '-', 'values', 'are', 'automatically', 'assigned', 'to', 'the', 'correct', 'section', 'of', 'the', 'QuantFigure'] | train | https://github.com/santosjorge/cufflinks/blob/ca1cbf93998dc793d0b1f8ac30fe1f2bd105f63a/cufflinks/quant_figure.py#L186-L203 |
8,206 | inveniosoftware-attic/invenio-knowledge | invenio_knowledge/api.py | get_kbs_info | def get_kbs_info(kbtype="", searchkbname=""):
"""A convenience method.
:param kbtype: type of kb -- get only kb's of this type
:param searchkbname: get only kb's where this sting appears in the name
"""
# query + order by
query = models.KnwKB.query.order_by(
models.KnwKB.name)
# fil... | python | def get_kbs_info(kbtype="", searchkbname=""):
"""A convenience method.
:param kbtype: type of kb -- get only kb's of this type
:param searchkbname: get only kb's where this sting appears in the name
"""
# query + order by
query = models.KnwKB.query.order_by(
models.KnwKB.name)
# fil... | ['def', 'get_kbs_info', '(', 'kbtype', '=', '""', ',', 'searchkbname', '=', '""', ')', ':', '# query + order by', 'query', '=', 'models', '.', 'KnwKB', '.', 'query', '.', 'order_by', '(', 'models', '.', 'KnwKB', '.', 'name', ')', '# filters', 'if', 'kbtype', ':', 'query', '=', 'query', '.', 'filter_by', '(', 'kbtype', ... | A convenience method.
:param kbtype: type of kb -- get only kb's of this type
:param searchkbname: get only kb's where this sting appears in the name | ['A', 'convenience', 'method', '.'] | train | https://github.com/inveniosoftware-attic/invenio-knowledge/blob/b31722dc14243ca8f626f8b3bce9718d0119de55/invenio_knowledge/api.py#L409-L424 |
8,207 | serge-sans-paille/pythran | pythran/types/conversion.py | pytype_to_ctype | def pytype_to_ctype(t):
""" Python -> pythonic type binding. """
if isinstance(t, List):
return 'pythonic::types::list<{0}>'.format(
pytype_to_ctype(t.__args__[0])
)
elif isinstance(t, Set):
return 'pythonic::types::set<{0}>'.format(
pytype_to_ctype(t.__args__... | python | def pytype_to_ctype(t):
""" Python -> pythonic type binding. """
if isinstance(t, List):
return 'pythonic::types::list<{0}>'.format(
pytype_to_ctype(t.__args__[0])
)
elif isinstance(t, Set):
return 'pythonic::types::set<{0}>'.format(
pytype_to_ctype(t.__args__... | ['def', 'pytype_to_ctype', '(', 't', ')', ':', 'if', 'isinstance', '(', 't', ',', 'List', ')', ':', 'return', "'pythonic::types::list<{0}>'", '.', 'format', '(', 'pytype_to_ctype', '(', 't', '.', '__args__', '[', '0', ']', ')', ')', 'elif', 'isinstance', '(', 't', ',', 'Set', ')', ':', 'return', "'pythonic::types::set<... | Python -> pythonic type binding. | ['Python', '-', '>', 'pythonic', 'type', 'binding', '.'] | train | https://github.com/serge-sans-paille/pythran/blob/7e1b5af2dddfabc50bd2a977f0178be269b349b5/pythran/types/conversion.py#L43-L92 |
8,208 | skyfielders/python-skyfield | skyfield/timelib.py | Time.utc_datetime_and_leap_second | def utc_datetime_and_leap_second(self):
"""Convert to a Python ``datetime`` in UTC, plus a leap second value.
Convert this time to a `datetime`_ object and a leap second::
dt, leap_second = t.utc_datetime_and_leap_second()
If the third-party `pytz`_ package is available, then its
... | python | def utc_datetime_and_leap_second(self):
"""Convert to a Python ``datetime`` in UTC, plus a leap second value.
Convert this time to a `datetime`_ object and a leap second::
dt, leap_second = t.utc_datetime_and_leap_second()
If the third-party `pytz`_ package is available, then its
... | ['def', 'utc_datetime_and_leap_second', '(', 'self', ')', ':', 'year', ',', 'month', ',', 'day', ',', 'hour', ',', 'minute', ',', 'second', '=', 'self', '.', '_utc_tuple', '(', '_half_millisecond', ')', 'second', ',', 'fraction', '=', 'divmod', '(', 'second', ',', '1.0', ')', 'second', '=', 'second', '.', 'astype', '('... | Convert to a Python ``datetime`` in UTC, plus a leap second value.
Convert this time to a `datetime`_ object and a leap second::
dt, leap_second = t.utc_datetime_and_leap_second()
If the third-party `pytz`_ package is available, then its
``utc`` timezone will be used as the timezo... | ['Convert', 'to', 'a', 'Python', 'datetime', 'in', 'UTC', 'plus', 'a', 'leap', 'second', 'value', '.'] | train | https://github.com/skyfielders/python-skyfield/blob/51d9e042e06457f6b1f2415296d50a38cb3a300f/skyfield/timelib.py#L425-L462 |
8,209 | davesque/django-rest-framework-simplejwt | rest_framework_simplejwt/authentication.py | JWTAuthentication.get_raw_token | def get_raw_token(self, header):
"""
Extracts an unvalidated JSON web token from the given "Authorization"
header value.
"""
parts = header.split()
if len(parts) == 0:
# Empty AUTHORIZATION header sent
return None
if parts[0] not in AUTH_... | python | def get_raw_token(self, header):
"""
Extracts an unvalidated JSON web token from the given "Authorization"
header value.
"""
parts = header.split()
if len(parts) == 0:
# Empty AUTHORIZATION header sent
return None
if parts[0] not in AUTH_... | ['def', 'get_raw_token', '(', 'self', ',', 'header', ')', ':', 'parts', '=', 'header', '.', 'split', '(', ')', 'if', 'len', '(', 'parts', ')', '==', '0', ':', '# Empty AUTHORIZATION header sent', 'return', 'None', 'if', 'parts', '[', '0', ']', 'not', 'in', 'AUTH_HEADER_TYPE_BYTES', ':', '# Assume the header does not co... | Extracts an unvalidated JSON web token from the given "Authorization"
header value. | ['Extracts', 'an', 'unvalidated', 'JSON', 'web', 'token', 'from', 'the', 'given', 'Authorization', 'header', 'value', '.'] | train | https://github.com/davesque/django-rest-framework-simplejwt/blob/d6084c595aefbf97865d15254b56017e710e8e47/rest_framework_simplejwt/authentication.py#L59-L80 |
8,210 | lsbardel/python-stdnet | stdnet/odm/struct.py | OrderedMixin.ipop_range | def ipop_range(self, start=0, stop=-1, callback=None, withscores=True):
'''pop a range from the :class:`OrderedMixin`'''
backend = self.backend
res = backend.structure(self).ipop_range(start, stop,
withscores=withscores)
if not callba... | python | def ipop_range(self, start=0, stop=-1, callback=None, withscores=True):
'''pop a range from the :class:`OrderedMixin`'''
backend = self.backend
res = backend.structure(self).ipop_range(start, stop,
withscores=withscores)
if not callba... | ['def', 'ipop_range', '(', 'self', ',', 'start', '=', '0', ',', 'stop', '=', '-', '1', ',', 'callback', '=', 'None', ',', 'withscores', '=', 'True', ')', ':', 'backend', '=', 'self', '.', 'backend', 'res', '=', 'backend', '.', 'structure', '(', 'self', ')', '.', 'ipop_range', '(', 'start', ',', 'stop', ',', 'withscores... | pop a range from the :class:`OrderedMixin` | ['pop', 'a', 'range', 'from', 'the', ':', 'class', ':', 'OrderedMixin'] | train | https://github.com/lsbardel/python-stdnet/blob/78db5320bdedc3f28c5e4f38cda13a4469e35db7/stdnet/odm/struct.py#L558-L565 |
8,211 | ray-project/ray | python/ray/profiling.py | Profiler._periodically_flush_profile_events | def _periodically_flush_profile_events(self):
"""Drivers run this as a thread to flush profile data in the
background."""
# Note(rkn): This is run on a background thread in the driver. It uses
# the raylet client. This should be ok because it doesn't read
# from the raylet client... | python | def _periodically_flush_profile_events(self):
"""Drivers run this as a thread to flush profile data in the
background."""
# Note(rkn): This is run on a background thread in the driver. It uses
# the raylet client. This should be ok because it doesn't read
# from the raylet client... | ['def', '_periodically_flush_profile_events', '(', 'self', ')', ':', '# Note(rkn): This is run on a background thread in the driver. It uses', "# the raylet client. This should be ok because it doesn't read", '# from the raylet client and we have the GIL here. However,', '# if either of those things changes, then we co... | Drivers run this as a thread to flush profile data in the
background. | ['Drivers', 'run', 'this', 'as', 'a', 'thread', 'to', 'flush', 'profile', 'data', 'in', 'the', 'background', '.'] | train | https://github.com/ray-project/ray/blob/4eade036a0505e244c976f36aaa2d64386b5129b/python/ray/profiling.py#L94-L110 |
8,212 | iotile/coretools | iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Environment.py | Base.ReplaceIxes | def ReplaceIxes(self, path, old_prefix, old_suffix, new_prefix, new_suffix):
"""
Replace old_prefix with new_prefix and old_suffix with new_suffix.
env - Environment used to interpolate variables.
path - the path that will be modified.
old_prefix - construction variable for the ... | python | def ReplaceIxes(self, path, old_prefix, old_suffix, new_prefix, new_suffix):
"""
Replace old_prefix with new_prefix and old_suffix with new_suffix.
env - Environment used to interpolate variables.
path - the path that will be modified.
old_prefix - construction variable for the ... | ['def', 'ReplaceIxes', '(', 'self', ',', 'path', ',', 'old_prefix', ',', 'old_suffix', ',', 'new_prefix', ',', 'new_suffix', ')', ':', 'old_prefix', '=', 'self', '.', 'subst', '(', "'$'", '+', 'old_prefix', ')', 'old_suffix', '=', 'self', '.', 'subst', '(', "'$'", '+', 'old_suffix', ')', 'new_prefix', '=', 'self', '.',... | Replace old_prefix with new_prefix and old_suffix with new_suffix.
env - Environment used to interpolate variables.
path - the path that will be modified.
old_prefix - construction variable for the old prefix.
old_suffix - construction variable for the old suffix.
new_prefix - c... | ['Replace', 'old_prefix', 'with', 'new_prefix', 'and', 'old_suffix', 'with', 'new_suffix', '.'] | train | https://github.com/iotile/coretools/blob/2d794f5f1346b841b0dcd16c9d284e9bf2f3c6ec/iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Environment.py#L1749-L1771 |
8,213 | ansibleplaybookbundle/ansible-playbook-bundle | src/apb/cli.py | subcmd_list_parser | def subcmd_list_parser(subcmd):
""" list subcommand """
subcmd.add_argument(
'--broker',
action='store',
dest='broker',
help=u'Route to the Ansible Service Broker'
)
subcmd.add_argument(
'--secure',
action='store_true',
dest='verify',
help=... | python | def subcmd_list_parser(subcmd):
""" list subcommand """
subcmd.add_argument(
'--broker',
action='store',
dest='broker',
help=u'Route to the Ansible Service Broker'
)
subcmd.add_argument(
'--secure',
action='store_true',
dest='verify',
help=... | ['def', 'subcmd_list_parser', '(', 'subcmd', ')', ':', 'subcmd', '.', 'add_argument', '(', "'--broker'", ',', 'action', '=', "'store'", ',', 'dest', '=', "'broker'", ',', 'help', '=', "u'Route to the Ansible Service Broker'", ')', 'subcmd', '.', 'add_argument', '(', "'--secure'", ',', 'action', '=', "'store_true'", ','... | list subcommand | ['list', 'subcommand'] | train | https://github.com/ansibleplaybookbundle/ansible-playbook-bundle/blob/585694be9b417f1a88354cbfe286bfd68c2c9494/src/apb/cli.py#L30-L85 |
8,214 | kwikteam/phy | phy/plot/transform.py | _glslify | def _glslify(r):
"""Transform a string or a n-tuple to a valid GLSL expression."""
if isinstance(r, string_types):
return r
else:
assert 2 <= len(r) <= 4
return 'vec{}({})'.format(len(r), ', '.join(map(str, r))) | python | def _glslify(r):
"""Transform a string or a n-tuple to a valid GLSL expression."""
if isinstance(r, string_types):
return r
else:
assert 2 <= len(r) <= 4
return 'vec{}({})'.format(len(r), ', '.join(map(str, r))) | ['def', '_glslify', '(', 'r', ')', ':', 'if', 'isinstance', '(', 'r', ',', 'string_types', ')', ':', 'return', 'r', 'else', ':', 'assert', '2', '<=', 'len', '(', 'r', ')', '<=', '4', 'return', "'vec{}({})'", '.', 'format', '(', 'len', '(', 'r', ')', ',', "', '", '.', 'join', '(', 'map', '(', 'str', ',', 'r', ')', ')', ... | Transform a string or a n-tuple to a valid GLSL expression. | ['Transform', 'a', 'string', 'or', 'a', 'n', '-', 'tuple', 'to', 'a', 'valid', 'GLSL', 'expression', '.'] | train | https://github.com/kwikteam/phy/blob/7e9313dc364304b7d2bd03b92938347343703003/phy/plot/transform.py#L48-L54 |
8,215 | rshk/python-libxdo | xdo/__init__.py | Xdo.wait_for_window_map_state | def wait_for_window_map_state(self, window, state):
"""
Wait for a window to have a specific map state.
State possibilities:
IsUnmapped - window is not displayed.
IsViewable - window is mapped and shown (though may be
clipped by windows on top of it)
... | python | def wait_for_window_map_state(self, window, state):
"""
Wait for a window to have a specific map state.
State possibilities:
IsUnmapped - window is not displayed.
IsViewable - window is mapped and shown (though may be
clipped by windows on top of it)
... | ['def', 'wait_for_window_map_state', '(', 'self', ',', 'window', ',', 'state', ')', ':', '_libxdo', '.', 'xdo_wait_for_window_map_state', '(', 'self', '.', '_xdo', ',', 'window', ',', 'state', ')'] | Wait for a window to have a specific map state.
State possibilities:
IsUnmapped - window is not displayed.
IsViewable - window is mapped and shown (though may be
clipped by windows on top of it)
IsUnviewable - window is mapped but a parent window is unmapped.
... | ['Wait', 'for', 'a', 'window', 'to', 'have', 'a', 'specific', 'map', 'state', '.'] | train | https://github.com/rshk/python-libxdo/blob/84cafa5943b005bc423edd28203a5266b3579ac3/xdo/__init__.py#L335-L348 |
8,216 | christophertbrown/bioscripts | ctbBio/genome_abundance.py | absolute_abundance | def absolute_abundance(coverage, total_bases):
"""
absolute abundance = (number of bases mapped to genome / total number of bases in sample) * 100
"""
absolute = {}
for genome in coverage:
absolute[genome] = []
index = 0
for calc in coverage[genome]:
bases = calc[0]
total = total_bases[index]
absolu... | python | def absolute_abundance(coverage, total_bases):
"""
absolute abundance = (number of bases mapped to genome / total number of bases in sample) * 100
"""
absolute = {}
for genome in coverage:
absolute[genome] = []
index = 0
for calc in coverage[genome]:
bases = calc[0]
total = total_bases[index]
absolu... | ['def', 'absolute_abundance', '(', 'coverage', ',', 'total_bases', ')', ':', 'absolute', '=', '{', '}', 'for', 'genome', 'in', 'coverage', ':', 'absolute', '[', 'genome', ']', '=', '[', ']', 'index', '=', '0', 'for', 'calc', 'in', 'coverage', '[', 'genome', ']', ':', 'bases', '=', 'calc', '[', '0', ']', 'total', '=', '... | absolute abundance = (number of bases mapped to genome / total number of bases in sample) * 100 | ['absolute', 'abundance', '=', '(', 'number', 'of', 'bases', 'mapped', 'to', 'genome', '/', 'total', 'number', 'of', 'bases', 'in', 'sample', ')', '*', '100'] | train | https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/genome_abundance.py#L45-L65 |
8,217 | usc-isi-i2/etk | etk/knowledge_graph.py | KnowledgeGraph.add_value | def add_value(self, field_name: str, value: object = None, json_path: str = None,
json_path_extraction: str = None, keep_empty: bool = False) -> None:
"""
Add a value to knowledge graph.
Input can either be a value or a json_path. If the input is json_path, the helper function ... | python | def add_value(self, field_name: str, value: object = None, json_path: str = None,
json_path_extraction: str = None, keep_empty: bool = False) -> None:
"""
Add a value to knowledge graph.
Input can either be a value or a json_path. If the input is json_path, the helper function ... | ['def', 'add_value', '(', 'self', ',', 'field_name', ':', 'str', ',', 'value', ':', 'object', '=', 'None', ',', 'json_path', ':', 'str', '=', 'None', ',', 'json_path_extraction', ':', 'str', '=', 'None', ',', 'keep_empty', ':', 'bool', '=', 'False', ')', '->', 'None', ':', 'def', 'validate', '(', 'v', ')', ':', 'if', '... | Add a value to knowledge graph.
Input can either be a value or a json_path. If the input is json_path, the helper function _add_doc_value is
called.
If the input is a value, then it is handled
Args:
field_name: str, the field name in the knowledge graph
value: th... | ['Add', 'a', 'value', 'to', 'knowledge', 'graph', '.', 'Input', 'can', 'either', 'be', 'a', 'value', 'or', 'a', 'json_path', '.', 'If', 'the', 'input', 'is', 'json_path', 'the', 'helper', 'function', '_add_doc_value', 'is', 'called', '.', 'If', 'the', 'input', 'is', 'a', 'value', 'then', 'it', 'is', 'handled'] | train | https://github.com/usc-isi-i2/etk/blob/aab077c984ea20f5e8ae33af622fe11d3c4df866/etk/knowledge_graph.py#L134-L194 |
8,218 | scott-maddox/openbandparams | src/openbandparams/iii_v_zinc_blende_alloy.py | IIIVZincBlendeAlloy.meff_SO | def meff_SO(self, **kwargs):
'''
Returns the split-off hole effective mass
calculated from Eg_Gamma(T), Delta_SO, Ep and F.
Interpolation of Eg_Gamma(T), Delta_SO, Ep and luttinger1, and
then calculation of meff_SO is recommended for alloys.
'''
Eg = self... | python | def meff_SO(self, **kwargs):
'''
Returns the split-off hole effective mass
calculated from Eg_Gamma(T), Delta_SO, Ep and F.
Interpolation of Eg_Gamma(T), Delta_SO, Ep and luttinger1, and
then calculation of meff_SO is recommended for alloys.
'''
Eg = self... | ['def', 'meff_SO', '(', 'self', ',', '*', '*', 'kwargs', ')', ':', 'Eg', '=', 'self', '.', 'Eg_Gamma', '(', '*', '*', 'kwargs', ')', 'Delta_SO', '=', 'self', '.', 'Delta_SO', '(', '*', '*', 'kwargs', ')', 'Ep', '=', 'self', '.', 'Ep', '(', '*', '*', 'kwargs', ')', 'luttinger1', '=', 'self', '.', 'luttinger1', '(', '*',... | Returns the split-off hole effective mass
calculated from Eg_Gamma(T), Delta_SO, Ep and F.
Interpolation of Eg_Gamma(T), Delta_SO, Ep and luttinger1, and
then calculation of meff_SO is recommended for alloys. | ['Returns', 'the', 'split', '-', 'off', 'hole', 'effective', 'mass', 'calculated', 'from', 'Eg_Gamma', '(', 'T', ')', 'Delta_SO', 'Ep', 'and', 'F', '.', 'Interpolation', 'of', 'Eg_Gamma', '(', 'T', ')', 'Delta_SO', 'Ep', 'and', 'luttinger1', 'and', 'then', 'calculation', 'of', 'meff_SO', 'is', 'recommended', 'for', 'al... | train | https://github.com/scott-maddox/openbandparams/blob/bc24e59187326bcb8948117434536082c9055777/src/openbandparams/iii_v_zinc_blende_alloy.py#L168-L180 |
8,219 | NiklasRosenstein-Python/nr-deprecated | nr/tools/versionupgrade.py | parse_config | def parse_config(filename):
"""
Parses a versionupgrade configuration file. Example:
tag v{VERSION}
branch v{VERSION}
message Prepare {VERSION} release
upgrade setup.py: version = '{VERSION}'
upgrade __init__.py:__version__ = '{VERSION}'
sub docs/changelog/v{VERSION}.md:# v{VER... | python | def parse_config(filename):
"""
Parses a versionupgrade configuration file. Example:
tag v{VERSION}
branch v{VERSION}
message Prepare {VERSION} release
upgrade setup.py: version = '{VERSION}'
upgrade __init__.py:__version__ = '{VERSION}'
sub docs/changelog/v{VERSION}.md:# v{VER... | ['def', 'parse_config', '(', 'filename', ')', ':', 'tag', '=', 'None', 'branch', '=', 'None', 'message', '=', "'Prepare {VERSION} release.'", 'upgrades', '=', '{', '}', 'subs', '=', '{', '}', 'with', 'open', '(', 'filename', ')', 'as', 'fp', ':', 'for', 'i', ',', 'line', 'in', 'enumerate', '(', 'fp', ')', ':', 'line', ... | Parses a versionupgrade configuration file. Example:
tag v{VERSION}
branch v{VERSION}
message Prepare {VERSION} release
upgrade setup.py: version = '{VERSION}'
upgrade __init__.py:__version__ = '{VERSION}'
sub docs/changelog/v{VERSION}.md:# v{VERSION} (unreleased):# v{VERSION} ({DA... | ['Parses', 'a', 'versionupgrade', 'configuration', 'file', '.', 'Example', ':'] | train | https://github.com/NiklasRosenstein-Python/nr-deprecated/blob/f9f8b89ea1b084841a8ab65784eaf68852686b2a/nr/tools/versionupgrade.py#L40-L102 |
8,220 | ampl/amplpy | amplpy/ampl.py | AMPL.getObjectives | def getObjectives(self):
"""
Get all the objectives declared.
"""
objectives = lock_and_call(
lambda: self._impl.getObjectives(),
self._lock
)
return EntityMap(objectives, Objective) | python | def getObjectives(self):
"""
Get all the objectives declared.
"""
objectives = lock_and_call(
lambda: self._impl.getObjectives(),
self._lock
)
return EntityMap(objectives, Objective) | ['def', 'getObjectives', '(', 'self', ')', ':', 'objectives', '=', 'lock_and_call', '(', 'lambda', ':', 'self', '.', '_impl', '.', 'getObjectives', '(', ')', ',', 'self', '.', '_lock', ')', 'return', 'EntityMap', '(', 'objectives', ',', 'Objective', ')'] | Get all the objectives declared. | ['Get', 'all', 'the', 'objectives', 'declared', '.'] | train | https://github.com/ampl/amplpy/blob/39df6954049a11a8f666aed26853259b4687099a/amplpy/ampl.py#L819-L827 |
8,221 | PmagPy/PmagPy | programs/demag_gui.py | Demag_GUI.convert_ages_to_calendar_year | def convert_ages_to_calendar_year(self, er_ages_rec):
"""
convert all age units to calendar year
Parameters
----------
er_ages_rec : Dict type object containing preferbly at least keys
'age', 'age_unit', and either 'age_range_high', 'age_range_low'
or 'ag... | python | def convert_ages_to_calendar_year(self, er_ages_rec):
"""
convert all age units to calendar year
Parameters
----------
er_ages_rec : Dict type object containing preferbly at least keys
'age', 'age_unit', and either 'age_range_high', 'age_range_low'
or 'ag... | ['def', 'convert_ages_to_calendar_year', '(', 'self', ',', 'er_ages_rec', ')', ':', 'if', '"age"', 'not', 'in', 'list', '(', 'er_ages_rec', '.', 'keys', '(', ')', ')', ':', 'return', '(', 'er_ages_rec', ')', 'if', '"age_unit"', 'not', 'in', 'list', '(', 'er_ages_rec', '.', 'keys', '(', ')', ')', ':', 'return', '(', 'er... | convert all age units to calendar year
Parameters
----------
er_ages_rec : Dict type object containing preferbly at least keys
'age', 'age_unit', and either 'age_range_high', 'age_range_low'
or 'age_sigma'
Returns
-------
er_ages_rec : Same dict ... | ['convert', 'all', 'age', 'units', 'to', 'calendar', 'year'] | train | https://github.com/PmagPy/PmagPy/blob/c7984f8809bf40fe112e53dcc311a33293b62d0b/programs/demag_gui.py#L2773-L2846 |
8,222 | kennknowles/python-jsonpath-rw | jsonpath_rw/jsonpath.py | DatumInContext.id_pseudopath | def id_pseudopath(self):
"""
Looks like a path, but with ids stuck in when available
"""
try:
pseudopath = Fields(str(self.value[auto_id_field]))
except (TypeError, AttributeError, KeyError): # This may not be all the interesting exceptions
pseudopath = se... | python | def id_pseudopath(self):
"""
Looks like a path, but with ids stuck in when available
"""
try:
pseudopath = Fields(str(self.value[auto_id_field]))
except (TypeError, AttributeError, KeyError): # This may not be all the interesting exceptions
pseudopath = se... | ['def', 'id_pseudopath', '(', 'self', ')', ':', 'try', ':', 'pseudopath', '=', 'Fields', '(', 'str', '(', 'self', '.', 'value', '[', 'auto_id_field', ']', ')', ')', 'except', '(', 'TypeError', ',', 'AttributeError', ',', 'KeyError', ')', ':', '# This may not be all the interesting exceptions', 'pseudopath', '=', 'self'... | Looks like a path, but with ids stuck in when available | ['Looks', 'like', 'a', 'path', 'but', 'with', 'ids', 'stuck', 'in', 'when', 'available'] | train | https://github.com/kennknowles/python-jsonpath-rw/blob/f615451d7b405e23e0f80b15cad03b1427b0256d/jsonpath_rw/jsonpath.py#L97-L109 |
8,223 | ejeschke/ginga | ginga/rv/plugins/PlotTable.py | PlotTable.plot_two_columns | def plot_two_columns(self, reset_xlimits=False, reset_ylimits=False):
"""Simple line plot for two selected columns."""
self.clear_plot()
if self.tab is None: # No table data to plot
return
plt_kw = {
'lw': self.settings.get('linewidth', 1),
'ls': se... | python | def plot_two_columns(self, reset_xlimits=False, reset_ylimits=False):
"""Simple line plot for two selected columns."""
self.clear_plot()
if self.tab is None: # No table data to plot
return
plt_kw = {
'lw': self.settings.get('linewidth', 1),
'ls': se... | ['def', 'plot_two_columns', '(', 'self', ',', 'reset_xlimits', '=', 'False', ',', 'reset_ylimits', '=', 'False', ')', ':', 'self', '.', 'clear_plot', '(', ')', 'if', 'self', '.', 'tab', 'is', 'None', ':', '# No table data to plot', 'return', 'plt_kw', '=', '{', "'lw'", ':', 'self', '.', 'settings', '.', 'get', '(', "'l... | Simple line plot for two selected columns. | ['Simple', 'line', 'plot', 'for', 'two', 'selected', 'columns', '.'] | train | https://github.com/ejeschke/ginga/blob/a78c893ec6f37a837de851947e9bb4625c597915/ginga/rv/plugins/PlotTable.py#L237-L274 |
8,224 | deepmind/sonnet | sonnet/examples/dataset_shakespeare.py | TinyShakespeareDataset._build | def _build(self):
"""Returns a tuple containing observation and target one-hot tensors."""
q = tf.FIFOQueue(
self._queue_capacity, [self._dtype, self._dtype],
shapes=[[self._num_steps, self._batch_size, self._vocab_size]]*2)
obs, target = tf.py_func(self._get_batch, [], [tf.int32, tf.int32])... | python | def _build(self):
"""Returns a tuple containing observation and target one-hot tensors."""
q = tf.FIFOQueue(
self._queue_capacity, [self._dtype, self._dtype],
shapes=[[self._num_steps, self._batch_size, self._vocab_size]]*2)
obs, target = tf.py_func(self._get_batch, [], [tf.int32, tf.int32])... | ['def', '_build', '(', 'self', ')', ':', 'q', '=', 'tf', '.', 'FIFOQueue', '(', 'self', '.', '_queue_capacity', ',', '[', 'self', '.', '_dtype', ',', 'self', '.', '_dtype', ']', ',', 'shapes', '=', '[', '[', 'self', '.', '_num_steps', ',', 'self', '.', '_batch_size', ',', 'self', '.', '_vocab_size', ']', ']', '*', '2',... | Returns a tuple containing observation and target one-hot tensors. | ['Returns', 'a', 'tuple', 'containing', 'observation', 'and', 'target', 'one', '-', 'hot', 'tensors', '.'] | train | https://github.com/deepmind/sonnet/blob/00612ca3178964d86b556e062694d808ff81fcca/sonnet/examples/dataset_shakespeare.py#L183-L194 |
8,225 | ArchiveTeam/wpull | wpull/warc/recorder.py | WARCRecorder.set_length_and_maybe_checksums | def set_length_and_maybe_checksums(self, record, payload_offset=None):
'''Set the content length and possibly the checksums.'''
if self._params.digests:
record.compute_checksum(payload_offset)
else:
record.set_content_length() | python | def set_length_and_maybe_checksums(self, record, payload_offset=None):
'''Set the content length and possibly the checksums.'''
if self._params.digests:
record.compute_checksum(payload_offset)
else:
record.set_content_length() | ['def', 'set_length_and_maybe_checksums', '(', 'self', ',', 'record', ',', 'payload_offset', '=', 'None', ')', ':', 'if', 'self', '.', '_params', '.', 'digests', ':', 'record', '.', 'compute_checksum', '(', 'payload_offset', ')', 'else', ':', 'record', '.', 'set_content_length', '(', ')'] | Set the content length and possibly the checksums. | ['Set', 'the', 'content', 'length', 'and', 'possibly', 'the', 'checksums', '.'] | train | https://github.com/ArchiveTeam/wpull/blob/ddf051aa3322479325ba20aa778cb2cb97606bf5/wpull/warc/recorder.py#L304-L309 |
8,226 | camptocamp/Studio | studio/controllers/datastores.py | DatastoresController.create | def create(self):
"""POST /datastores: Create a new item."""
# url('datastores')
content = request.environ['wsgi.input'].read(int(request.environ['CONTENT_LENGTH']))
content = content.decode('utf8')
content = simplejson.loads(content)
new = DataStore(content['name'],
... | python | def create(self):
"""POST /datastores: Create a new item."""
# url('datastores')
content = request.environ['wsgi.input'].read(int(request.environ['CONTENT_LENGTH']))
content = content.decode('utf8')
content = simplejson.loads(content)
new = DataStore(content['name'],
... | ['def', 'create', '(', 'self', ')', ':', "# url('datastores')", 'content', '=', 'request', '.', 'environ', '[', "'wsgi.input'", ']', '.', 'read', '(', 'int', '(', 'request', '.', 'environ', '[', "'CONTENT_LENGTH'", ']', ')', ')', 'content', '=', 'content', '.', 'decode', '(', "'utf8'", ')', 'content', '=', 'simplejson'... | POST /datastores: Create a new item. | ['POST', '/', 'datastores', ':', 'Create', 'a', 'new', 'item', '.'] | train | https://github.com/camptocamp/Studio/blob/43cb7298434fb606b15136801b79b03571a2f27e/studio/controllers/datastores.py#L53-L64 |
8,227 | PhracturedBlue/asterisk_mbox | asterisk_mbox/__init__.py | Client.get_cdr | def get_cdr(self, start=0, count=-1, **kwargs):
"""Request range of CDR messages"""
sha = encode_to_sha("{:d},{:d}".format(start, count))
return self._queue_msg({'cmd': cmd.CMD_MESSAGE_CDR,
'sha': sha}, **kwargs) | python | def get_cdr(self, start=0, count=-1, **kwargs):
"""Request range of CDR messages"""
sha = encode_to_sha("{:d},{:d}".format(start, count))
return self._queue_msg({'cmd': cmd.CMD_MESSAGE_CDR,
'sha': sha}, **kwargs) | ['def', 'get_cdr', '(', 'self', ',', 'start', '=', '0', ',', 'count', '=', '-', '1', ',', '*', '*', 'kwargs', ')', ':', 'sha', '=', 'encode_to_sha', '(', '"{:d},{:d}"', '.', 'format', '(', 'start', ',', 'count', ')', ')', 'return', 'self', '.', '_queue_msg', '(', '{', "'cmd'", ':', 'cmd', '.', 'CMD_MESSAGE_CDR', ',', "... | Request range of CDR messages | ['Request', 'range', 'of', 'CDR', 'messages'] | train | https://github.com/PhracturedBlue/asterisk_mbox/blob/275de1e71ed05c6acff1a5fa87f754f4d385a372/asterisk_mbox/__init__.py#L231-L235 |
8,228 | Stranger6667/postmarker | postmarker/models/emails.py | EmailBatch.send | def send(self, **extra):
"""
Sends email batch.
:return: Information about sent emails.
:rtype: `list`
"""
emails = self.as_dict(**extra)
responses = [self._manager._send_batch(*batch) for batch in chunks(emails, self.MAX_SIZE)]
return sum(responses, []) | python | def send(self, **extra):
"""
Sends email batch.
:return: Information about sent emails.
:rtype: `list`
"""
emails = self.as_dict(**extra)
responses = [self._manager._send_batch(*batch) for batch in chunks(emails, self.MAX_SIZE)]
return sum(responses, []) | ['def', 'send', '(', 'self', ',', '*', '*', 'extra', ')', ':', 'emails', '=', 'self', '.', 'as_dict', '(', '*', '*', 'extra', ')', 'responses', '=', '[', 'self', '.', '_manager', '.', '_send_batch', '(', '*', 'batch', ')', 'for', 'batch', 'in', 'chunks', '(', 'emails', ',', 'self', '.', 'MAX_SIZE', ')', ']', 'return', ... | Sends email batch.
:return: Information about sent emails.
:rtype: `list` | ['Sends', 'email', 'batch', '.'] | train | https://github.com/Stranger6667/postmarker/blob/013224ab1761e95c488c7d2701e6fa83f3108d94/postmarker/models/emails.py#L256-L265 |
8,229 | BernardFW/bernard | src/bernard/cli/_live_reload.py | start_child | async def start_child():
"""
Start the child process that will look for changes in modules.
"""
logger.info('Started to watch for code changes')
loop = asyncio.get_event_loop()
watcher = aionotify.Watcher()
flags = (
aionotify.Flags.MODIFY |
aionotify.Flags.DELETE |
... | python | async def start_child():
"""
Start the child process that will look for changes in modules.
"""
logger.info('Started to watch for code changes')
loop = asyncio.get_event_loop()
watcher = aionotify.Watcher()
flags = (
aionotify.Flags.MODIFY |
aionotify.Flags.DELETE |
... | ['async', 'def', 'start_child', '(', ')', ':', 'logger', '.', 'info', '(', "'Started to watch for code changes'", ')', 'loop', '=', 'asyncio', '.', 'get_event_loop', '(', ')', 'watcher', '=', 'aionotify', '.', 'Watcher', '(', ')', 'flags', '=', '(', 'aionotify', '.', 'Flags', '.', 'MODIFY', '|', 'aionotify', '.', 'Flag... | Start the child process that will look for changes in modules. | ['Start', 'the', 'child', 'process', 'that', 'will', 'look', 'for', 'changes', 'in', 'modules', '.'] | train | https://github.com/BernardFW/bernard/blob/9c55703e5ffe5717c9fa39793df59dbfa5b4c5ab/src/bernard/cli/_live_reload.py#L83-L120 |
8,230 | CamDavidsonPilon/lifetimes | lifetimes/plotting.py | plot_expected_repeat_purchases | def plot_expected_repeat_purchases(
model,
title="Expected Number of Repeat Purchases per Customer",
xlabel="Time Since First Purchase",
ax=None,
label=None,
**kwargs
):
"""
Plot expected repeat purchases on calibration period .
Parameters
----------
model: lifetimes model
... | python | def plot_expected_repeat_purchases(
model,
title="Expected Number of Repeat Purchases per Customer",
xlabel="Time Since First Purchase",
ax=None,
label=None,
**kwargs
):
"""
Plot expected repeat purchases on calibration period .
Parameters
----------
model: lifetimes model
... | ['def', 'plot_expected_repeat_purchases', '(', 'model', ',', 'title', '=', '"Expected Number of Repeat Purchases per Customer"', ',', 'xlabel', '=', '"Time Since First Purchase"', ',', 'ax', '=', 'None', ',', 'label', '=', 'None', ',', '*', '*', 'kwargs', ')', ':', 'from', 'matplotlib', 'import', 'pyplot', 'as', 'plt',... | Plot expected repeat purchases on calibration period .
Parameters
----------
model: lifetimes model
A fitted lifetimes model.
max_frequency: int, optional
The maximum frequency to plot.
title: str, optional
Figure title
xlabel: str, optional
Figure xlabel
ax:... | ['Plot', 'expected', 'repeat', 'purchases', 'on', 'calibration', 'period', '.'] | train | https://github.com/CamDavidsonPilon/lifetimes/blob/f926308bc03c17c1d12fead729de43885cf13321/lifetimes/plotting.py#L270-L326 |
8,231 | LonamiWebs/Telethon | telethon/tl/custom/button.py | Button.request_location | def request_location(cls, text, *,
resize=None, single_use=None, selective=None):
"""
Creates a new button that will request
the user's location upon being clicked.
``resize``, ``single_use`` and ``selective`` are documented in `text`.
"""
return... | python | def request_location(cls, text, *,
resize=None, single_use=None, selective=None):
"""
Creates a new button that will request
the user's location upon being clicked.
``resize``, ``single_use`` and ``selective`` are documented in `text`.
"""
return... | ['def', 'request_location', '(', 'cls', ',', 'text', ',', '*', ',', 'resize', '=', 'None', ',', 'single_use', '=', 'None', ',', 'selective', '=', 'None', ')', ':', 'return', 'cls', '(', 'types', '.', 'KeyboardButtonRequestGeoLocation', '(', 'text', ')', ',', 'resize', '=', 'resize', ',', 'single_use', '=', 'single_use'... | Creates a new button that will request
the user's location upon being clicked.
``resize``, ``single_use`` and ``selective`` are documented in `text`. | ['Creates', 'a', 'new', 'button', 'that', 'will', 'request', 'the', 'user', 's', 'location', 'upon', 'being', 'clicked', '.'] | train | https://github.com/LonamiWebs/Telethon/blob/1ead9757d366b58c1e0567cddb0196e20f1a445f/telethon/tl/custom/button.py#L122-L131 |
8,232 | djgagne/hagelslag | hagelslag/processing/ObjectMatcher.py | centroid_distance | def centroid_distance(item_a, time_a, item_b, time_b, max_value):
"""
Euclidean distance between the centroids of item_a and item_b.
Args:
item_a: STObject from the first set in ObjectMatcher
time_a: Time integer being evaluated
item_b: STObject from the second set in ObjectMatcher
... | python | def centroid_distance(item_a, time_a, item_b, time_b, max_value):
"""
Euclidean distance between the centroids of item_a and item_b.
Args:
item_a: STObject from the first set in ObjectMatcher
time_a: Time integer being evaluated
item_b: STObject from the second set in ObjectMatcher
... | ['def', 'centroid_distance', '(', 'item_a', ',', 'time_a', ',', 'item_b', ',', 'time_b', ',', 'max_value', ')', ':', 'ax', ',', 'ay', '=', 'item_a', '.', 'center_of_mass', '(', 'time_a', ')', 'bx', ',', 'by', '=', 'item_b', '.', 'center_of_mass', '(', 'time_b', ')', 'return', 'np', '.', 'minimum', '(', 'np', '.', 'sqrt... | Euclidean distance between the centroids of item_a and item_b.
Args:
item_a: STObject from the first set in ObjectMatcher
time_a: Time integer being evaluated
item_b: STObject from the second set in ObjectMatcher
time_b: Time integer being evaluated
max_value: Maximum distan... | ['Euclidean', 'distance', 'between', 'the', 'centroids', 'of', 'item_a', 'and', 'item_b', '.'] | train | https://github.com/djgagne/hagelslag/blob/6fb6c3df90bf4867e13a97d3460b14471d107df1/hagelslag/processing/ObjectMatcher.py#L246-L262 |
8,233 | quasipedia/swaggery | swaggery/checker.py | Checker.checks | def checks(self):
'''Return the list of all check methods.'''
condition = lambda a: a.startswith('check_')
return (getattr(self, a) for a in dir(self) if condition(a)) | python | def checks(self):
'''Return the list of all check methods.'''
condition = lambda a: a.startswith('check_')
return (getattr(self, a) for a in dir(self) if condition(a)) | ['def', 'checks', '(', 'self', ')', ':', 'condition', '=', 'lambda', 'a', ':', 'a', '.', 'startswith', '(', "'check_'", ')', 'return', '(', 'getattr', '(', 'self', ',', 'a', ')', 'for', 'a', 'in', 'dir', '(', 'self', ')', 'if', 'condition', '(', 'a', ')', ')'] | Return the list of all check methods. | ['Return', 'the', 'list', 'of', 'all', 'check', 'methods', '.'] | train | https://github.com/quasipedia/swaggery/blob/89a2e1b2bebbc511c781c9e63972f65aef73cc2f/swaggery/checker.py#L43-L46 |
8,234 | Julius2342/pyvlx | pyvlx/nodes.py | Nodes.add | def add(self, node):
"""Add Node, replace existing node if node with node_id is present."""
if not isinstance(node, Node):
raise TypeError()
for i, j in enumerate(self.__nodes):
if j.node_id == node.node_id:
self.__nodes[i] = node
return
... | python | def add(self, node):
"""Add Node, replace existing node if node with node_id is present."""
if not isinstance(node, Node):
raise TypeError()
for i, j in enumerate(self.__nodes):
if j.node_id == node.node_id:
self.__nodes[i] = node
return
... | ['def', 'add', '(', 'self', ',', 'node', ')', ':', 'if', 'not', 'isinstance', '(', 'node', ',', 'Node', ')', ':', 'raise', 'TypeError', '(', ')', 'for', 'i', ',', 'j', 'in', 'enumerate', '(', 'self', '.', '__nodes', ')', ':', 'if', 'j', '.', 'node_id', '==', 'node', '.', 'node_id', ':', 'self', '.', '__nodes', '[', 'i'... | Add Node, replace existing node if node with node_id is present. | ['Add', 'Node', 'replace', 'existing', 'node', 'if', 'node', 'with', 'node_id', 'is', 'present', '.'] | train | https://github.com/Julius2342/pyvlx/blob/ee78e1324bcb1be5b8d1a9d05ab5496b72eae848/pyvlx/nodes.py#L51-L59 |
8,235 | SavinaRoja/OpenAccess_EPUB | src/openaccess_epub/publisher/plos.py | PLoS.move_back_boxed_texts | def move_back_boxed_texts(self):
"""
The only intended use for this function is to patch a problem seen in
at least one PLoS article (journal.pgen.0020002). This will move any
<boxed-text> elements over to the receiving element, which is probably
the main body.
"""
... | python | def move_back_boxed_texts(self):
"""
The only intended use for this function is to patch a problem seen in
at least one PLoS article (journal.pgen.0020002). This will move any
<boxed-text> elements over to the receiving element, which is probably
the main body.
"""
... | ['def', 'move_back_boxed_texts', '(', 'self', ')', ':', 'body', '=', 'self', '.', 'main', '.', 'getroot', '(', ')', '.', 'find', '(', "'body'", ')', 'back', '=', 'self', '.', 'article', '.', 'root', '.', 'find', '(', "'back'", ')', 'if', 'back', 'is', 'None', ':', 'return', 'boxed_texts', '=', 'back', '.', 'xpath', '('... | The only intended use for this function is to patch a problem seen in
at least one PLoS article (journal.pgen.0020002). This will move any
<boxed-text> elements over to the receiving element, which is probably
the main body. | ['The', 'only', 'intended', 'use', 'for', 'this', 'function', 'is', 'to', 'patch', 'a', 'problem', 'seen', 'in', 'at', 'least', 'one', 'PLoS', 'article', '(', 'journal', '.', 'pgen', '.', '0020002', ')', '.', 'This', 'will', 'move', 'any', '<boxed', '-', 'text', '>', 'elements', 'over', 'to', 'the', 'receiving', 'eleme... | train | https://github.com/SavinaRoja/OpenAccess_EPUB/blob/6b77ba30b7394fd003920e7a7957bca963a90656/src/openaccess_epub/publisher/plos.py#L682-L695 |
8,236 | mozilla/mozilla-django-oidc | mozilla_django_oidc/middleware.py | SessionRefresh.is_refreshable_url | def is_refreshable_url(self, request):
"""Takes a request and returns whether it triggers a refresh examination
:arg HttpRequest request:
:returns: boolean
"""
# Do not attempt to refresh the session if the OIDC backend is not used
backend_session = request.session.get... | python | def is_refreshable_url(self, request):
"""Takes a request and returns whether it triggers a refresh examination
:arg HttpRequest request:
:returns: boolean
"""
# Do not attempt to refresh the session if the OIDC backend is not used
backend_session = request.session.get... | ['def', 'is_refreshable_url', '(', 'self', ',', 'request', ')', ':', '# Do not attempt to refresh the session if the OIDC backend is not used', 'backend_session', '=', 'request', '.', 'session', '.', 'get', '(', 'BACKEND_SESSION_KEY', ')', 'is_oidc_enabled', '=', 'True', 'if', 'backend_session', ':', 'auth_backend', '=... | Takes a request and returns whether it triggers a refresh examination
:arg HttpRequest request:
:returns: boolean | ['Takes', 'a', 'request', 'and', 'returns', 'whether', 'it', 'triggers', 'a', 'refresh', 'examination'] | train | https://github.com/mozilla/mozilla-django-oidc/blob/e780130deacccbafc85a92f48d1407e042f5f955/mozilla_django_oidc/middleware.py#L67-L87 |
8,237 | bukun/TorCMS | torcms/handlers/collect_handler.py | CollectHandler.add_or_update | def add_or_update(self, app_id):
'''
Add or update the category.
'''
logger.info('Collect info: user-{0}, uid-{1}'.format(self.userinfo.uid, app_id))
MCollect.add_or_update(self.userinfo.uid, app_id)
out_dic = {'success': True}
return json.dump(out_dic, self) | python | def add_or_update(self, app_id):
'''
Add or update the category.
'''
logger.info('Collect info: user-{0}, uid-{1}'.format(self.userinfo.uid, app_id))
MCollect.add_or_update(self.userinfo.uid, app_id)
out_dic = {'success': True}
return json.dump(out_dic, self) | ['def', 'add_or_update', '(', 'self', ',', 'app_id', ')', ':', 'logger', '.', 'info', '(', "'Collect info: user-{0}, uid-{1}'", '.', 'format', '(', 'self', '.', 'userinfo', '.', 'uid', ',', 'app_id', ')', ')', 'MCollect', '.', 'add_or_update', '(', 'self', '.', 'userinfo', '.', 'uid', ',', 'app_id', ')', 'out_dic', '='... | Add or update the category. | ['Add', 'or', 'update', 'the', 'category', '.'] | train | https://github.com/bukun/TorCMS/blob/6567c7fe2604a1d646d4570c017840958630ed2b/torcms/handlers/collect_handler.py#L44-L51 |
8,238 | lsbardel/python-stdnet | stdnet/odm/related.py | makeMany2ManyRelatedManager | def makeMany2ManyRelatedManager(formodel, name_relmodel, name_formodel):
'''formodel is the model which the manager .'''
class _Many2ManyRelatedManager(Many2ManyRelatedManager):
pass
_Many2ManyRelatedManager.formodel = formodel
_Many2ManyRelatedManager.name_relmodel = name_relmodel
... | python | def makeMany2ManyRelatedManager(formodel, name_relmodel, name_formodel):
'''formodel is the model which the manager .'''
class _Many2ManyRelatedManager(Many2ManyRelatedManager):
pass
_Many2ManyRelatedManager.formodel = formodel
_Many2ManyRelatedManager.name_relmodel = name_relmodel
... | ['def', 'makeMany2ManyRelatedManager', '(', 'formodel', ',', 'name_relmodel', ',', 'name_formodel', ')', ':', 'class', '_Many2ManyRelatedManager', '(', 'Many2ManyRelatedManager', ')', ':', 'pass', '_Many2ManyRelatedManager', '.', 'formodel', '=', 'formodel', '_Many2ManyRelatedManager', '.', 'name_relmodel', '=', 'name_... | formodel is the model which the manager . | ['formodel', 'is', 'the', 'model', 'which', 'the', 'manager', '.'] | train | https://github.com/lsbardel/python-stdnet/blob/78db5320bdedc3f28c5e4f38cda13a4469e35db7/stdnet/odm/related.py#L265-L274 |
8,239 | mabuchilab/QNET | src/qnet/algebra/core/state_algebra.py | BasisKet.next | def next(self, n=1):
"""Move up by `n` steps in the Hilbert space::
>>> hs = LocalSpace('tls', basis=('g', 'e'))
>>> ascii(BasisKet('g', hs=hs).next())
'|e>^(tls)'
>>> ascii(BasisKet(0, hs=hs).next())
'|e>^(tls)'
We can also go multiple step... | python | def next(self, n=1):
"""Move up by `n` steps in the Hilbert space::
>>> hs = LocalSpace('tls', basis=('g', 'e'))
>>> ascii(BasisKet('g', hs=hs).next())
'|e>^(tls)'
>>> ascii(BasisKet(0, hs=hs).next())
'|e>^(tls)'
We can also go multiple step... | ['def', 'next', '(', 'self', ',', 'n', '=', '1', ')', ':', 'if', 'isinstance', '(', 'self', '.', 'label', ',', 'SymbolicLabelBase', ')', ':', 'next_label', '=', 'self', '.', 'space', '.', 'next_basis_label_or_index', '(', 'self', '.', 'label', ',', 'n', ')', 'return', 'BasisKet', '(', 'next_label', ',', 'hs', '=', 'sel... | Move up by `n` steps in the Hilbert space::
>>> hs = LocalSpace('tls', basis=('g', 'e'))
>>> ascii(BasisKet('g', hs=hs).next())
'|e>^(tls)'
>>> ascii(BasisKet(0, hs=hs).next())
'|e>^(tls)'
We can also go multiple steps:
>>> hs = LocalS... | ['Move', 'up', 'by', 'n', 'steps', 'in', 'the', 'Hilbert', 'space', '::'] | train | https://github.com/mabuchilab/QNET/blob/cc20d26dad78691d34c67173e5cd67dcac94208a/src/qnet/algebra/core/state_algebra.py#L322-L353 |
8,240 | spyder-ide/spyder-kernels | spyder_kernels/console/kernel.py | SpyderKernel.load_data | def load_data(self, filename, ext):
"""Load data from filename"""
from spyder_kernels.utils.iofuncs import iofunctions
from spyder_kernels.utils.misc import fix_reference_name
glbs = self._mglobals()
load_func = iofunctions.load_funcs[ext]
data, error_message = load_fun... | python | def load_data(self, filename, ext):
"""Load data from filename"""
from spyder_kernels.utils.iofuncs import iofunctions
from spyder_kernels.utils.misc import fix_reference_name
glbs = self._mglobals()
load_func = iofunctions.load_funcs[ext]
data, error_message = load_fun... | ['def', 'load_data', '(', 'self', ',', 'filename', ',', 'ext', ')', ':', 'from', 'spyder_kernels', '.', 'utils', '.', 'iofuncs', 'import', 'iofunctions', 'from', 'spyder_kernels', '.', 'utils', '.', 'misc', 'import', 'fix_reference_name', 'glbs', '=', 'self', '.', '_mglobals', '(', ')', 'load_func', '=', 'iofunctions',... | Load data from filename | ['Load', 'data', 'from', 'filename'] | train | https://github.com/spyder-ide/spyder-kernels/blob/2c5b36cdb797b8aba77bc406ca96f5e079c4aaca/spyder_kernels/console/kernel.py#L194-L217 |
8,241 | jlmadurga/permabots | permabots/views/hooks/telegram_hook.py | TelegramHookView.post | def post(self, request, hook_id):
"""
Process Telegram webhook.
1. Serialize Telegram message
2. Get an enabled Telegram bot
3. Create :class:`Update <permabots.models.telegram_api.Update>`
5. Delay processing to a task
6. Response provid... | python | def post(self, request, hook_id):
"""
Process Telegram webhook.
1. Serialize Telegram message
2. Get an enabled Telegram bot
3. Create :class:`Update <permabots.models.telegram_api.Update>`
5. Delay processing to a task
6. Response provid... | ['def', 'post', '(', 'self', ',', 'request', ',', 'hook_id', ')', ':', 'serializer', '=', 'UpdateSerializer', '(', 'data', '=', 'request', '.', 'data', ')', 'if', 'serializer', '.', 'is_valid', '(', ')', ':', 'try', ':', 'bot', '=', 'caching', '.', 'get_or_set', '(', 'TelegramBot', ',', 'hook_id', ')', 'except', 'Teleg... | Process Telegram webhook.
1. Serialize Telegram message
2. Get an enabled Telegram bot
3. Create :class:`Update <permabots.models.telegram_api.Update>`
5. Delay processing to a task
6. Response provider | ['Process', 'Telegram', 'webhook', '.', '1', '.', 'Serialize', 'Telegram', 'message', '2', '.', 'Get', 'an', 'enabled', 'Telegram', 'bot', '3', '.', 'Create', ':', 'class', ':', 'Update', '<permabots', '.', 'models', '.', 'telegram_api', '.', 'Update', '>', '5', '.', 'Delay', 'processing', 'to', 'a', 'task', '6', '.', ... | train | https://github.com/jlmadurga/permabots/blob/781a91702529a23fe7bc2aa84c5d88e961412466/permabots/views/hooks/telegram_hook.py#L87-L121 |
8,242 | saltstack/salt | salt/modules/elasticsearch.py | index_template_exists | def index_template_exists(name, hosts=None, profile=None):
'''
Return a boolean indicating whether given index template exists
name
Index template name
CLI example::
salt myminion elasticsearch.index_template_exists testindex_templ
'''
es = _get_instance(hosts, profile)
tr... | python | def index_template_exists(name, hosts=None, profile=None):
'''
Return a boolean indicating whether given index template exists
name
Index template name
CLI example::
salt myminion elasticsearch.index_template_exists testindex_templ
'''
es = _get_instance(hosts, profile)
tr... | ['def', 'index_template_exists', '(', 'name', ',', 'hosts', '=', 'None', ',', 'profile', '=', 'None', ')', ':', 'es', '=', '_get_instance', '(', 'hosts', ',', 'profile', ')', 'try', ':', 'return', 'es', '.', 'indices', '.', 'exists_template', '(', 'name', '=', 'name', ')', 'except', 'elasticsearch', '.', 'TransportErro... | Return a boolean indicating whether given index template exists
name
Index template name
CLI example::
salt myminion elasticsearch.index_template_exists testindex_templ | ['Return', 'a', 'boolean', 'indicating', 'whether', 'given', 'index', 'template', 'exists'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/elasticsearch.py#L776-L791 |
8,243 | sckott/pygbif | pygbif/occurrences/get.py | get_verbatim | def get_verbatim(key, **kwargs):
'''
Gets a verbatim occurrence record without any interpretation
:param key: [int] A GBIF occurrence key
:return: A dictionary, of results
Usage::
from pygbif import occurrences
occurrences.get_verbatim(key = 1258202889)
occurrences.get_ve... | python | def get_verbatim(key, **kwargs):
'''
Gets a verbatim occurrence record without any interpretation
:param key: [int] A GBIF occurrence key
:return: A dictionary, of results
Usage::
from pygbif import occurrences
occurrences.get_verbatim(key = 1258202889)
occurrences.get_ve... | ['def', 'get_verbatim', '(', 'key', ',', '*', '*', 'kwargs', ')', ':', 'url', '=', 'gbif_baseurl', '+', "'occurrence/'", '+', 'str', '(', 'key', ')', '+', "'/verbatim'", 'out', '=', 'gbif_GET', '(', 'url', ',', '{', '}', ',', '*', '*', 'kwargs', ')', 'return', 'out'] | Gets a verbatim occurrence record without any interpretation
:param key: [int] A GBIF occurrence key
:return: A dictionary, of results
Usage::
from pygbif import occurrences
occurrences.get_verbatim(key = 1258202889)
occurrences.get_verbatim(key = 1227768771)
occurrences.... | ['Gets', 'a', 'verbatim', 'occurrence', 'record', 'without', 'any', 'interpretation'] | train | https://github.com/sckott/pygbif/blob/bf54f2920bc46d97d7e2e1b0c8059e5878f3c5ab/pygbif/occurrences/get.py#L22-L39 |
8,244 | google/grr | grr/server/grr_response_server/aff4.py | AttributeExpression.Compile | def Compile(self, filter_implemention):
"""Returns the data_store filter implementation from the attribute."""
return self.operator_method(self.attribute_obj, filter_implemention,
*self.args) | python | def Compile(self, filter_implemention):
"""Returns the data_store filter implementation from the attribute."""
return self.operator_method(self.attribute_obj, filter_implemention,
*self.args) | ['def', 'Compile', '(', 'self', ',', 'filter_implemention', ')', ':', 'return', 'self', '.', 'operator_method', '(', 'self', '.', 'attribute_obj', ',', 'filter_implemention', ',', '*', 'self', '.', 'args', ')'] | Returns the data_store filter implementation from the attribute. | ['Returns', 'the', 'data_store', 'filter', 'implementation', 'from', 'the', 'attribute', '.'] | train | https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/server/grr_response_server/aff4.py#L2373-L2376 |
8,245 | kennethreitz/records | records.py | Record.dataset | def dataset(self):
"""A Tablib Dataset containing the row."""
data = tablib.Dataset()
data.headers = self.keys()
row = _reduce_datetimes(self.values())
data.append(row)
return data | python | def dataset(self):
"""A Tablib Dataset containing the row."""
data = tablib.Dataset()
data.headers = self.keys()
row = _reduce_datetimes(self.values())
data.append(row)
return data | ['def', 'dataset', '(', 'self', ')', ':', 'data', '=', 'tablib', '.', 'Dataset', '(', ')', 'data', '.', 'headers', '=', 'self', '.', 'keys', '(', ')', 'row', '=', '_reduce_datetimes', '(', 'self', '.', 'values', '(', ')', ')', 'data', '.', 'append', '(', 'row', ')', 'return', 'data'] | A Tablib Dataset containing the row. | ['A', 'Tablib', 'Dataset', 'containing', 'the', 'row', '.'] | train | https://github.com/kennethreitz/records/blob/ecd857266c5e7830d657cbe0196816314790563b/records.py#L88-L96 |
8,246 | DeV1doR/aioethereum | aioethereum/utils.py | add_0x | def add_0x(string):
"""Add 0x to string at start.
"""
if isinstance(string, bytes):
string = string.decode('utf-8')
return '0x' + str(string) | python | def add_0x(string):
"""Add 0x to string at start.
"""
if isinstance(string, bytes):
string = string.decode('utf-8')
return '0x' + str(string) | ['def', 'add_0x', '(', 'string', ')', ':', 'if', 'isinstance', '(', 'string', ',', 'bytes', ')', ':', 'string', '=', 'string', '.', 'decode', '(', "'utf-8'", ')', 'return', "'0x'", '+', 'str', '(', 'string', ')'] | Add 0x to string at start. | ['Add', '0x', 'to', 'string', 'at', 'start', '.'] | train | https://github.com/DeV1doR/aioethereum/blob/85eb46550d862b3ccc309914ea871ca1c7b42157/aioethereum/utils.py#L4-L9 |
8,247 | rochacbruno/python-pagseguro | examples/flask/flask_seguro/controllers/main/__init__.py | add_to_cart | def add_to_cart(item_id):
""" Cart with Product """
cart = Cart(session['cart'])
if cart.change_item(item_id, 'add'):
session['cart'] = cart.to_dict()
return list_products() | python | def add_to_cart(item_id):
""" Cart with Product """
cart = Cart(session['cart'])
if cart.change_item(item_id, 'add'):
session['cart'] = cart.to_dict()
return list_products() | ['def', 'add_to_cart', '(', 'item_id', ')', ':', 'cart', '=', 'Cart', '(', 'session', '[', "'cart'", ']', ')', 'if', 'cart', '.', 'change_item', '(', 'item_id', ',', "'add'", ')', ':', 'session', '[', "'cart'", ']', '=', 'cart', '.', 'to_dict', '(', ')', 'return', 'list_products', '(', ')'] | Cart with Product | ['Cart', 'with', 'Product'] | train | https://github.com/rochacbruno/python-pagseguro/blob/18a9ca3301783cb323e838574b59f9ddffa9a593/examples/flask/flask_seguro/controllers/main/__init__.py#L44-L49 |
8,248 | funkybob/antfarm | antfarm/response.py | Response.build_headers | def build_headers(self):
'''
Return the list of headers as two-tuples
'''
if not 'Content-Type' in self.headers:
content_type = self.content_type
if self.encoding != DEFAULT_ENCODING:
content_type += '; charset=%s' % self.encoding
self.... | python | def build_headers(self):
'''
Return the list of headers as two-tuples
'''
if not 'Content-Type' in self.headers:
content_type = self.content_type
if self.encoding != DEFAULT_ENCODING:
content_type += '; charset=%s' % self.encoding
self.... | ['def', 'build_headers', '(', 'self', ')', ':', 'if', 'not', "'Content-Type'", 'in', 'self', '.', 'headers', ':', 'content_type', '=', 'self', '.', 'content_type', 'if', 'self', '.', 'encoding', '!=', 'DEFAULT_ENCODING', ':', 'content_type', '+=', "'; charset=%s'", '%', 'self', '.', 'encoding', 'self', '.', 'headers', ... | Return the list of headers as two-tuples | ['Return', 'the', 'list', 'of', 'headers', 'as', 'two', '-', 'tuples'] | train | https://github.com/funkybob/antfarm/blob/40a7cc450eba09a280b7bc8f7c68a807b0177c62/antfarm/response.py#L90-L106 |
8,249 | rix0rrr/gcl | gcl/doc.py | sort_members | def sort_members(tup, names):
"""Return two pairs of members, scalar and tuple members.
The scalars will be sorted s.t. the unbound members are at the top.
"""
scalars, tuples = partition(lambda x: not is_tuple_node(tup.member[x].value), names)
unbound, bound = partition(lambda x: tup.member[x].value.is_unbo... | python | def sort_members(tup, names):
"""Return two pairs of members, scalar and tuple members.
The scalars will be sorted s.t. the unbound members are at the top.
"""
scalars, tuples = partition(lambda x: not is_tuple_node(tup.member[x].value), names)
unbound, bound = partition(lambda x: tup.member[x].value.is_unbo... | ['def', 'sort_members', '(', 'tup', ',', 'names', ')', ':', 'scalars', ',', 'tuples', '=', 'partition', '(', 'lambda', 'x', ':', 'not', 'is_tuple_node', '(', 'tup', '.', 'member', '[', 'x', ']', '.', 'value', ')', ',', 'names', ')', 'unbound', ',', 'bound', '=', 'partition', '(', 'lambda', 'x', ':', 'tup', '.', 'member... | Return two pairs of members, scalar and tuple members.
The scalars will be sorted s.t. the unbound members are at the top. | ['Return', 'two', 'pairs', 'of', 'members', 'scalar', 'and', 'tuple', 'members', '.'] | train | https://github.com/rix0rrr/gcl/blob/4e3bccc978a9c60aaaffd20f6f291c4d23775cdf/gcl/doc.py#L229-L236 |
8,250 | saltstack/salt | salt/modules/virt.py | pool_build | def pool_build(name, **kwargs):
'''
Build a defined libvirt storage pool.
:param name: libvirt storage pool name
:param connection: libvirt connection URI, overriding defaults
:param username: username to connect with, overriding defaults
:param password: password to connect with, overriding de... | python | def pool_build(name, **kwargs):
'''
Build a defined libvirt storage pool.
:param name: libvirt storage pool name
:param connection: libvirt connection URI, overriding defaults
:param username: username to connect with, overriding defaults
:param password: password to connect with, overriding de... | ['def', 'pool_build', '(', 'name', ',', '*', '*', 'kwargs', ')', ':', 'conn', '=', '__get_conn', '(', '*', '*', 'kwargs', ')', 'try', ':', 'pool', '=', 'conn', '.', 'storagePoolLookupByName', '(', 'name', ')', 'return', 'not', 'bool', '(', 'pool', '.', 'build', '(', ')', ')', 'finally', ':', 'conn', '.', 'close', '(', ... | Build a defined libvirt storage pool.
:param name: libvirt storage pool name
:param connection: libvirt connection URI, overriding defaults
:param username: username to connect with, overriding defaults
:param password: password to connect with, overriding defaults
.. versionadded:: 2019.2.0
... | ['Build', 'a', 'defined', 'libvirt', 'storage', 'pool', '.'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/virt.py#L5085-L5107 |
8,251 | wavycloud/pyboto3 | pyboto3/cloudformation.py | create_stack | def create_stack(StackName=None, TemplateBody=None, TemplateURL=None, Parameters=None, DisableRollback=None, TimeoutInMinutes=None, NotificationARNs=None, Capabilities=None, ResourceTypes=None, RoleARN=None, OnFailure=None, StackPolicyBody=None, StackPolicyURL=None, Tags=None, ClientRequestToken=None):
"""
Crea... | python | def create_stack(StackName=None, TemplateBody=None, TemplateURL=None, Parameters=None, DisableRollback=None, TimeoutInMinutes=None, NotificationARNs=None, Capabilities=None, ResourceTypes=None, RoleARN=None, OnFailure=None, StackPolicyBody=None, StackPolicyURL=None, Tags=None, ClientRequestToken=None):
"""
Crea... | ['def', 'create_stack', '(', 'StackName', '=', 'None', ',', 'TemplateBody', '=', 'None', ',', 'TemplateURL', '=', 'None', ',', 'Parameters', '=', 'None', ',', 'DisableRollback', '=', 'None', ',', 'TimeoutInMinutes', '=', 'None', ',', 'NotificationARNs', '=', 'None', ',', 'Capabilities', '=', 'None', ',', 'ResourceTypes... | Creates a stack as specified in the template. After the call completes successfully, the stack creation starts. You can check the status of the stack via the DescribeStacks API.
See also: AWS API Documentation
:example: response = client.create_stack(
StackName='string',
TemplateBody=... | ['Creates', 'a', 'stack', 'as', 'specified', 'in', 'the', 'template', '.', 'After', 'the', 'call', 'completes', 'successfully', 'the', 'stack', 'creation', 'starts', '.', 'You', 'can', 'check', 'the', 'status', 'of', 'the', 'stack', 'via', 'the', 'DescribeStacks', 'API', '.', 'See', 'also', ':', 'AWS', 'API', 'Document... | train | https://github.com/wavycloud/pyboto3/blob/924957ccf994303713a4eed90b775ff2ab95b2e5/pyboto3/cloudformation.py#L255-L386 |
8,252 | pandas-dev/pandas | pandas/core/dtypes/cast.py | maybe_convert_platform | def maybe_convert_platform(values):
""" try to do platform conversion, allow ndarray or list here """
if isinstance(values, (list, tuple)):
values = construct_1d_object_array_from_listlike(list(values))
if getattr(values, 'dtype', None) == np.object_:
if hasattr(values, '_values'):
... | python | def maybe_convert_platform(values):
""" try to do platform conversion, allow ndarray or list here """
if isinstance(values, (list, tuple)):
values = construct_1d_object_array_from_listlike(list(values))
if getattr(values, 'dtype', None) == np.object_:
if hasattr(values, '_values'):
... | ['def', 'maybe_convert_platform', '(', 'values', ')', ':', 'if', 'isinstance', '(', 'values', ',', '(', 'list', ',', 'tuple', ')', ')', ':', 'values', '=', 'construct_1d_object_array_from_listlike', '(', 'list', '(', 'values', ')', ')', 'if', 'getattr', '(', 'values', ',', "'dtype'", ',', 'None', ')', '==', 'np', '.', ... | try to do platform conversion, allow ndarray or list here | ['try', 'to', 'do', 'platform', 'conversion', 'allow', 'ndarray', 'or', 'list', 'here'] | train | https://github.com/pandas-dev/pandas/blob/9feb3ad92cc0397a04b665803a49299ee7aa1037/pandas/core/dtypes/cast.py#L35-L45 |
8,253 | softlayer/softlayer-python | SoftLayer/CLI/block/replication/failover.py | cli | def cli(env, volume_id, replicant_id, immediate):
"""Failover a block volume to the given replicant volume."""
block_storage_manager = SoftLayer.BlockStorageManager(env.client)
success = block_storage_manager.failover_to_replicant(
volume_id,
replicant_id,
immediate
)
if su... | python | def cli(env, volume_id, replicant_id, immediate):
"""Failover a block volume to the given replicant volume."""
block_storage_manager = SoftLayer.BlockStorageManager(env.client)
success = block_storage_manager.failover_to_replicant(
volume_id,
replicant_id,
immediate
)
if su... | ['def', 'cli', '(', 'env', ',', 'volume_id', ',', 'replicant_id', ',', 'immediate', ')', ':', 'block_storage_manager', '=', 'SoftLayer', '.', 'BlockStorageManager', '(', 'env', '.', 'client', ')', 'success', '=', 'block_storage_manager', '.', 'failover_to_replicant', '(', 'volume_id', ',', 'replicant_id', ',', 'immedia... | Failover a block volume to the given replicant volume. | ['Failover', 'a', 'block', 'volume', 'to', 'the', 'given', 'replicant', 'volume', '.'] | train | https://github.com/softlayer/softlayer-python/blob/9f181be08cc3668353b05a6de0cb324f52cff6fa/SoftLayer/CLI/block/replication/failover.py#L17-L30 |
8,254 | dschep/lambda-decorators | lambda_decorators.py | on_exception | def on_exception(func):
"""
Run a function when a handler thows an exception. It's return value is
returned to AWS.
Usage::
>>> # to create a reusable decorator
>>> @on_exception
... def handle_errors(exception):
... print(exception)
... return {'statusC... | python | def on_exception(func):
"""
Run a function when a handler thows an exception. It's return value is
returned to AWS.
Usage::
>>> # to create a reusable decorator
>>> @on_exception
... def handle_errors(exception):
... print(exception)
... return {'statusC... | ['def', 'on_exception', '(', 'func', ')', ':', 'class', 'OnExceptionDecorator', '(', 'LambdaDecorator', ')', ':', 'def', 'on_exception', '(', 'self', ',', 'exception', ')', ':', 'return', 'func', '(', 'exception', ')', 'return', 'OnExceptionDecorator'] | Run a function when a handler thows an exception. It's return value is
returned to AWS.
Usage::
>>> # to create a reusable decorator
>>> @on_exception
... def handle_errors(exception):
... print(exception)
... return {'statusCode': 500, 'body': 'uh oh'}
... | ['Run', 'a', 'function', 'when', 'a', 'handler', 'thows', 'an', 'exception', '.', 'It', 's', 'return', 'value', 'is', 'returned', 'to', 'AWS', '.'] | train | https://github.com/dschep/lambda-decorators/blob/9195914c8afe26843de9968d96dae6a89f061e8a/lambda_decorators.py#L292-L321 |
8,255 | RI-imaging/qpformat | qpformat/file_formats/dataset.py | SeriesData.saveh5 | def saveh5(self, h5file, qpi_slice=None, series_slice=None,
time_interval=None, count=None, max_count=None):
"""Save the data set as an hdf5 file (qpimage.QPSeries format)
Parameters
----------
h5file: str, pathlib.Path, or h5py.Group
Where to store the series... | python | def saveh5(self, h5file, qpi_slice=None, series_slice=None,
time_interval=None, count=None, max_count=None):
"""Save the data set as an hdf5 file (qpimage.QPSeries format)
Parameters
----------
h5file: str, pathlib.Path, or h5py.Group
Where to store the series... | ['def', 'saveh5', '(', 'self', ',', 'h5file', ',', 'qpi_slice', '=', 'None', ',', 'series_slice', '=', 'None', ',', 'time_interval', '=', 'None', ',', 'count', '=', 'None', ',', 'max_count', '=', 'None', ')', ':', '# set up slice to export', 'if', 'series_slice', 'is', 'None', ':', 'sl', '=', 'range', '(', 'len', '(', ... | Save the data set as an hdf5 file (qpimage.QPSeries format)
Parameters
----------
h5file: str, pathlib.Path, or h5py.Group
Where to store the series data
qpi_slice: tuple of (slice, slice)
If not None, only store a slice of each QPImage
in `h5file`. A... | ['Save', 'the', 'data', 'set', 'as', 'an', 'hdf5', 'file', '(', 'qpimage', '.', 'QPSeries', 'format', ')'] | train | https://github.com/RI-imaging/qpformat/blob/364e29d7d9e8b9f1d7a4a25c753d1baf9d73d5eb/qpformat/file_formats/dataset.py#L211-L291 |
8,256 | lsbardel/python-stdnet | stdnet/utils/jsontools.py | addmul_number_dicts | def addmul_number_dicts(series):
'''Multiply dictionaries by a numeric values and add them together.
:parameter series: a tuple of two elements tuples. Each serie is of the form::
(weight,dictionary)
where ``weight`` is a number and ``dictionary`` is a dictionary with
numeric values.
:parameter s... | python | def addmul_number_dicts(series):
'''Multiply dictionaries by a numeric values and add them together.
:parameter series: a tuple of two elements tuples. Each serie is of the form::
(weight,dictionary)
where ``weight`` is a number and ``dictionary`` is a dictionary with
numeric values.
:parameter s... | ['def', 'addmul_number_dicts', '(', 'series', ')', ':', 'if', 'not', 'series', ':', 'return', 'vtype', '=', 'value_type', '(', '(', 's', '[', '1', ']', 'for', 's', 'in', 'series', ')', ')', 'if', 'vtype', '==', '1', ':', 'return', 'sum', '(', '(', 'weight', '*', 'float', '(', 'd', ')', 'for', 'weight', ',', 'd', 'in', ... | Multiply dictionaries by a numeric values and add them together.
:parameter series: a tuple of two elements tuples. Each serie is of the form::
(weight,dictionary)
where ``weight`` is a number and ``dictionary`` is a dictionary with
numeric values.
:parameter skip: optional list of field names to ski... | ['Multiply', 'dictionaries', 'by', 'a', 'numeric', 'values', 'and', 'add', 'them', 'together', '.'] | train | https://github.com/lsbardel/python-stdnet/blob/78db5320bdedc3f28c5e4f38cda13a4469e35db7/stdnet/utils/jsontools.py#L201-L229 |
8,257 | apache/incubator-mxnet | example/fcn-xs/init_fcnxs.py | init_from_fcnxs | def init_from_fcnxs(ctx, fcnxs_symbol, fcnxs_args_from, fcnxs_auxs_from):
""" use zero initialization for better convergence, because it tends to oputut 0,
and the label 0 stands for background, which may occupy most size of one image.
"""
fcnxs_args = fcnxs_args_from.copy()
fcnxs_auxs = fcnxs_auxs_... | python | def init_from_fcnxs(ctx, fcnxs_symbol, fcnxs_args_from, fcnxs_auxs_from):
""" use zero initialization for better convergence, because it tends to oputut 0,
and the label 0 stands for background, which may occupy most size of one image.
"""
fcnxs_args = fcnxs_args_from.copy()
fcnxs_auxs = fcnxs_auxs_... | ['def', 'init_from_fcnxs', '(', 'ctx', ',', 'fcnxs_symbol', ',', 'fcnxs_args_from', ',', 'fcnxs_auxs_from', ')', ':', 'fcnxs_args', '=', 'fcnxs_args_from', '.', 'copy', '(', ')', 'fcnxs_auxs', '=', 'fcnxs_auxs_from', '.', 'copy', '(', ')', 'for', 'k', ',', 'v', 'in', 'fcnxs_args', '.', 'items', '(', ')', ':', 'if', '('... | use zero initialization for better convergence, because it tends to oputut 0,
and the label 0 stands for background, which may occupy most size of one image. | ['use', 'zero', 'initialization', 'for', 'better', 'convergence', 'because', 'it', 'tends', 'to', 'oputut', '0', 'and', 'the', 'label', '0', 'stands', 'for', 'background', 'which', 'may', 'occupy', 'most', 'size', 'of', 'one', 'image', '.'] | train | https://github.com/apache/incubator-mxnet/blob/1af29e9c060a4c7d60eeaacba32afdb9a7775ba7/example/fcn-xs/init_fcnxs.py#L65-L106 |
8,258 | summa-tx/riemann | riemann/tx/sprout.py | SproutTx._sighash_prep | def _sighash_prep(self, index, script):
'''
SproutTx, int, byte-like -> SproutTx
Sighashes suck
Performs the sighash setup described here:
https://en.bitcoin.it/wiki/OP_CHECKSIG#How_it_works
https://bitcoin.stackexchange.com/questions/3374/how-to-redeem-a-basic-tx
... | python | def _sighash_prep(self, index, script):
'''
SproutTx, int, byte-like -> SproutTx
Sighashes suck
Performs the sighash setup described here:
https://en.bitcoin.it/wiki/OP_CHECKSIG#How_it_works
https://bitcoin.stackexchange.com/questions/3374/how-to-redeem-a-basic-tx
... | ['def', '_sighash_prep', '(', 'self', ',', 'index', ',', 'script', ')', ':', 'if', 'len', '(', 'self', '.', 'tx_ins', ')', '==', '0', ':', 'return', 'self', '.', 'copy', '(', 'joinsplit_sig', '=', "b''", ')', '# 0 out scripts in tx_ins', 'copy_tx_ins', '=', '[', 'tx_in', '.', 'copy', '(', 'stack_script', '=', "b''", ',... | SproutTx, int, byte-like -> SproutTx
Sighashes suck
Performs the sighash setup described here:
https://en.bitcoin.it/wiki/OP_CHECKSIG#How_it_works
https://bitcoin.stackexchange.com/questions/3374/how-to-redeem-a-basic-tx
We save on complexity by refusing to support OP_CODESEPARAT... | ['SproutTx', 'int', 'byte', '-', 'like', '-', '>', 'SproutTx', 'Sighashes', 'suck', 'Performs', 'the', 'sighash', 'setup', 'described', 'here', ':', 'https', ':', '//', 'en', '.', 'bitcoin', '.', 'it', '/', 'wiki', '/', 'OP_CHECKSIG#How_it_works', 'https', ':', '//', 'bitcoin', '.', 'stackexchange', '.', 'com', '/', 'q... | train | https://github.com/summa-tx/riemann/blob/04ae336dfd4007ceaed748daadc91cc32fa278ec/riemann/tx/sprout.py#L219-L240 |
8,259 | itamarst/eliot | eliot/journald.py | sd_journal_send | def sd_journal_send(**kwargs):
"""
Send a message to the journald log.
@param kwargs: Mapping between field names to values, both as bytes.
@raise IOError: If the operation failed.
"""
# The function uses printf formatting, so we need to quote
# percentages.
fields = [
_ffi.new... | python | def sd_journal_send(**kwargs):
"""
Send a message to the journald log.
@param kwargs: Mapping between field names to values, both as bytes.
@raise IOError: If the operation failed.
"""
# The function uses printf formatting, so we need to quote
# percentages.
fields = [
_ffi.new... | ['def', 'sd_journal_send', '(', '*', '*', 'kwargs', ')', ':', '# The function uses printf formatting, so we need to quote', '# percentages.', 'fields', '=', '[', '_ffi', '.', 'new', '(', '"char[]"', ',', 'key', '.', 'encode', '(', '"ascii"', ')', '+', "b'='", '+', 'value', '.', 'replace', '(', 'b"%"', ',', 'b"%%"', ')'... | Send a message to the journald log.
@param kwargs: Mapping between field names to values, both as bytes.
@raise IOError: If the operation failed. | ['Send', 'a', 'message', 'to', 'the', 'journald', 'log', '.'] | train | https://github.com/itamarst/eliot/blob/c03c96520c5492fadfc438b4b0f6336e2785ba2d/eliot/journald.py#L28-L45 |
8,260 | facetoe/zenpy | zenpy/lib/api.py | VariantApi.show | def show(self, item, variant):
"""
Show a variant.
:param item: Item object or id
:param variant: Variant object or id
:return:
"""
url = self._build_url(self.endpoint.show(item, variant))
return self._get(url) | python | def show(self, item, variant):
"""
Show a variant.
:param item: Item object or id
:param variant: Variant object or id
:return:
"""
url = self._build_url(self.endpoint.show(item, variant))
return self._get(url) | ['def', 'show', '(', 'self', ',', 'item', ',', 'variant', ')', ':', 'url', '=', 'self', '.', '_build_url', '(', 'self', '.', 'endpoint', '.', 'show', '(', 'item', ',', 'variant', ')', ')', 'return', 'self', '.', '_get', '(', 'url', ')'] | Show a variant.
:param item: Item object or id
:param variant: Variant object or id
:return: | ['Show', 'a', 'variant', '.'] | train | https://github.com/facetoe/zenpy/blob/34c54c7e408b9ed01604ddf8b3422204c8bf31ea/zenpy/lib/api.py#L1273-L1282 |
8,261 | kejbaly2/metrique | metrique/result.py | Result.unfinished_objects | def unfinished_objects(self):
'''
Leaves only versions of those objects that has some version with
`_end == None` or with `_end > right cutoff`.
'''
mask = self._end_isnull
if self._rbound is not None:
mask = mask | (self._end > self._rbound)
oids = se... | python | def unfinished_objects(self):
'''
Leaves only versions of those objects that has some version with
`_end == None` or with `_end > right cutoff`.
'''
mask = self._end_isnull
if self._rbound is not None:
mask = mask | (self._end > self._rbound)
oids = se... | ['def', 'unfinished_objects', '(', 'self', ')', ':', 'mask', '=', 'self', '.', '_end_isnull', 'if', 'self', '.', '_rbound', 'is', 'not', 'None', ':', 'mask', '=', 'mask', '|', '(', 'self', '.', '_end', '>', 'self', '.', '_rbound', ')', 'oids', '=', 'set', '(', 'self', '[', 'mask', ']', '.', '_oid', '.', 'tolist', '(', ... | Leaves only versions of those objects that has some version with
`_end == None` or with `_end > right cutoff`. | ['Leaves', 'only', 'versions', 'of', 'those', 'objects', 'that', 'has', 'some', 'version', 'with', '_end', '==', 'None', 'or', 'with', '_end', '>', 'right', 'cutoff', '.'] | train | https://github.com/kejbaly2/metrique/blob/a10b076097441b7dde687949139f702f5c1e1b35/metrique/result.py#L313-L322 |
8,262 | fp12/achallonge | challonge/participant.py | Participant.get_next_opponent | async def get_next_opponent(self):
""" Get the opponent of the potential next match. See :func:`get_next_match`
|methcoro|
Raises:
APIException
"""
next_match = await self.get_next_match()
if next_match is not None:
opponent_id = next_match.play... | python | async def get_next_opponent(self):
""" Get the opponent of the potential next match. See :func:`get_next_match`
|methcoro|
Raises:
APIException
"""
next_match = await self.get_next_match()
if next_match is not None:
opponent_id = next_match.play... | ['async', 'def', 'get_next_opponent', '(', 'self', ')', ':', 'next_match', '=', 'await', 'self', '.', 'get_next_match', '(', ')', 'if', 'next_match', 'is', 'not', 'None', ':', 'opponent_id', '=', 'next_match', '.', 'player1_id', 'if', 'next_match', '.', 'player2_id', '==', 'self', '.', '_id', 'else', 'next_match', '.',... | Get the opponent of the potential next match. See :func:`get_next_match`
|methcoro|
Raises:
APIException | ['Get', 'the', 'opponent', 'of', 'the', 'potential', 'next', 'match', '.', 'See', ':', 'func', ':', 'get_next_match'] | train | https://github.com/fp12/achallonge/blob/25780b3c48b66400a50ff9f884e4287afd4c89e4/challonge/participant.py#L194-L207 |
8,263 | spotify/snakebite | snakebite/client.py | Client.copyToLocal | def copyToLocal(self, paths, dst, check_crc=False):
''' Copy files that match the file source pattern
to the local name. Source is kept. When copying multiple,
files, the destination must be a directory.
:param paths: Paths to copy
:type paths: list of strings
:param d... | python | def copyToLocal(self, paths, dst, check_crc=False):
''' Copy files that match the file source pattern
to the local name. Source is kept. When copying multiple,
files, the destination must be a directory.
:param paths: Paths to copy
:type paths: list of strings
:param d... | ['def', 'copyToLocal', '(', 'self', ',', 'paths', ',', 'dst', ',', 'check_crc', '=', 'False', ')', ':', 'if', 'not', 'isinstance', '(', 'paths', ',', 'list', ')', ':', 'raise', 'InvalidInputException', '(', '"Paths should be a list"', ')', 'if', 'not', 'paths', ':', 'raise', 'InvalidInputException', '(', '"copyToLocal:... | Copy files that match the file source pattern
to the local name. Source is kept. When copying multiple,
files, the destination must be a directory.
:param paths: Paths to copy
:type paths: list of strings
:param dst: Destination path
:type dst: string
:param ch... | ['Copy', 'files', 'that', 'match', 'the', 'file', 'source', 'pattern', 'to', 'the', 'local', 'name', '.', 'Source', 'is', 'kept', '.', 'When', 'copying', 'multiple', 'files', 'the', 'destination', 'must', 'be', 'a', 'directory', '.'] | train | https://github.com/spotify/snakebite/blob/6a456e6100b0c1be66cc1f7f9d7f50494f369da3/snakebite/client.py#L707-L734 |
8,264 | molmod/molmod | molmod/examples/003_internal_coordinates/c_ff_hessian.py | ForceField.hessian | def hessian(self, coordinates):
"""Compute the force-field Hessian for the given coordinates.
Argument:
| ``coordinates`` -- A numpy array with the Cartesian atom
coordinates, with shape (N,3).
Returns:
| ``hessian`` -- A numpy arr... | python | def hessian(self, coordinates):
"""Compute the force-field Hessian for the given coordinates.
Argument:
| ``coordinates`` -- A numpy array with the Cartesian atom
coordinates, with shape (N,3).
Returns:
| ``hessian`` -- A numpy arr... | ['def', 'hessian', '(', 'self', ',', 'coordinates', ')', ':', '# N3 is 3 times the number of atoms.', 'N3', '=', 'coordinates', '.', 'size', '# Start with a zero hessian.', 'hessian', '=', 'numpy', '.', 'zeros', '(', '(', 'N3', ',', 'N3', ')', ',', 'float', ')', '# Add the contribution of each term.', 'for', 'term', 'i... | Compute the force-field Hessian for the given coordinates.
Argument:
| ``coordinates`` -- A numpy array with the Cartesian atom
coordinates, with shape (N,3).
Returns:
| ``hessian`` -- A numpy array with the Hessian, with shape (3*N,
... | ['Compute', 'the', 'force', '-', 'field', 'Hessian', 'for', 'the', 'given', 'coordinates', '.'] | train | https://github.com/molmod/molmod/blob/a7b5b4364ed514ad4c465856c05b5eda1cb561e0/molmod/examples/003_internal_coordinates/c_ff_hessian.py#L85-L103 |
8,265 | pyviz/holoviews | holoviews/plotting/bokeh/element.py | OverlayPlot.update_frame | def update_frame(self, key, ranges=None, element=None):
"""
Update the internal state of the Plot to represent the given
key tuple (where integers represent frames). Returns this
state.
"""
reused = isinstance(self.hmap, DynamicMap) and self.overlaid
if not reused... | python | def update_frame(self, key, ranges=None, element=None):
"""
Update the internal state of the Plot to represent the given
key tuple (where integers represent frames). Returns this
state.
"""
reused = isinstance(self.hmap, DynamicMap) and self.overlaid
if not reused... | ['def', 'update_frame', '(', 'self', ',', 'key', ',', 'ranges', '=', 'None', ',', 'element', '=', 'None', ')', ':', 'reused', '=', 'isinstance', '(', 'self', '.', 'hmap', ',', 'DynamicMap', ')', 'and', 'self', '.', 'overlaid', 'if', 'not', 'reused', 'and', 'element', 'is', 'None', ':', 'element', '=', 'self', '.', '_ge... | Update the internal state of the Plot to represent the given
key tuple (where integers represent frames). Returns this
state. | ['Update', 'the', 'internal', 'state', 'of', 'the', 'Plot', 'to', 'represent', 'the', 'given', 'key', 'tuple', '(', 'where', 'integers', 'represent', 'frames', ')', '.', 'Returns', 'this', 'state', '.'] | train | https://github.com/pyviz/holoviews/blob/ae0dd2f3de448b0ca5e9065aabd6ef8d84c7e655/holoviews/plotting/bokeh/element.py#L2106-L2180 |
8,266 | ForensicArtifacts/artifacts | artifacts/registry.py | ArtifactDefinitionsRegistry.ReadFromFile | def ReadFromFile(self, artifacts_reader, filename):
"""Reads artifact definitions into the registry from a file.
Args:
artifacts_reader (ArtifactsReader): an artifacts reader.
filename (str): name of the file to read from.
"""
for artifact_definition in artifacts_reader.ReadFile(filename):
... | python | def ReadFromFile(self, artifacts_reader, filename):
"""Reads artifact definitions into the registry from a file.
Args:
artifacts_reader (ArtifactsReader): an artifacts reader.
filename (str): name of the file to read from.
"""
for artifact_definition in artifacts_reader.ReadFile(filename):
... | ['def', 'ReadFromFile', '(', 'self', ',', 'artifacts_reader', ',', 'filename', ')', ':', 'for', 'artifact_definition', 'in', 'artifacts_reader', '.', 'ReadFile', '(', 'filename', ')', ':', 'self', '.', 'RegisterDefinition', '(', 'artifact_definition', ')'] | Reads artifact definitions into the registry from a file.
Args:
artifacts_reader (ArtifactsReader): an artifacts reader.
filename (str): name of the file to read from. | ['Reads', 'artifact', 'definitions', 'into', 'the', 'registry', 'from', 'a', 'file', '.'] | train | https://github.com/ForensicArtifacts/artifacts/blob/044a63bfb4448af33d085c69066c80f9505ae7ca/artifacts/registry.py#L200-L208 |
8,267 | fabioz/PyDev.Debugger | third_party/pep8/lib2to3/lib2to3/pytree.py | WildcardPattern.optimize | def optimize(self):
"""Optimize certain stacked wildcard patterns."""
subpattern = None
if (self.content is not None and
len(self.content) == 1 and len(self.content[0]) == 1):
subpattern = self.content[0][0]
if self.min == 1 and self.max == 1:
if self.... | python | def optimize(self):
"""Optimize certain stacked wildcard patterns."""
subpattern = None
if (self.content is not None and
len(self.content) == 1 and len(self.content[0]) == 1):
subpattern = self.content[0][0]
if self.min == 1 and self.max == 1:
if self.... | ['def', 'optimize', '(', 'self', ')', ':', 'subpattern', '=', 'None', 'if', '(', 'self', '.', 'content', 'is', 'not', 'None', 'and', 'len', '(', 'self', '.', 'content', ')', '==', '1', 'and', 'len', '(', 'self', '.', 'content', '[', '0', ']', ')', '==', '1', ')', ':', 'subpattern', '=', 'self', '.', 'content', '[', '0'... | Optimize certain stacked wildcard patterns. | ['Optimize', 'certain', 'stacked', 'wildcard', 'patterns', '.'] | train | https://github.com/fabioz/PyDev.Debugger/blob/ed9c4307662a5593b8a7f1f3389ecd0e79b8c503/third_party/pep8/lib2to3/lib2to3/pytree.py#L688-L705 |
8,268 | ionelmc/python-hunter | src/hunter/tracer.py | Tracer.stop | def stop(self):
"""
Stop tracing. Reinstalls the :ref:`hunter.Tracer.previous` tracer.
"""
if self._handler is not None:
sys.settrace(self._previous)
self._handler = self._previous = None
if self.threading_support is None or self.threading_support:
... | python | def stop(self):
"""
Stop tracing. Reinstalls the :ref:`hunter.Tracer.previous` tracer.
"""
if self._handler is not None:
sys.settrace(self._previous)
self._handler = self._previous = None
if self.threading_support is None or self.threading_support:
... | ['def', 'stop', '(', 'self', ')', ':', 'if', 'self', '.', '_handler', 'is', 'not', 'None', ':', 'sys', '.', 'settrace', '(', 'self', '.', '_previous', ')', 'self', '.', '_handler', '=', 'self', '.', '_previous', '=', 'None', 'if', 'self', '.', 'threading_support', 'is', 'None', 'or', 'self', '.', 'threading_support', '... | Stop tracing. Reinstalls the :ref:`hunter.Tracer.previous` tracer. | ['Stop', 'tracing', '.', 'Reinstalls', 'the', ':', 'ref', ':', 'hunter', '.', 'Tracer', '.', 'previous', 'tracer', '.'] | train | https://github.com/ionelmc/python-hunter/blob/b3a1310b0593d2c6b6ef430883843896e17d6a81/src/hunter/tracer.py#L98-L107 |
8,269 | materialsproject/pymatgen | pymatgen/analysis/surface_analysis.py | SurfaceEnergyPlotter.stable_u_range_dict | def stable_u_range_dict(self, chempot_range, ref_delu, no_doped=True,
no_clean=False, delu_dict={}, miller_index=(),
dmu_at_0=False, return_se_dict=False):
"""
Creates a dictionary where each entry is a key pointing to a
chemical potential ... | python | def stable_u_range_dict(self, chempot_range, ref_delu, no_doped=True,
no_clean=False, delu_dict={}, miller_index=(),
dmu_at_0=False, return_se_dict=False):
"""
Creates a dictionary where each entry is a key pointing to a
chemical potential ... | ['def', 'stable_u_range_dict', '(', 'self', ',', 'chempot_range', ',', 'ref_delu', ',', 'no_doped', '=', 'True', ',', 'no_clean', '=', 'False', ',', 'delu_dict', '=', '{', '}', ',', 'miller_index', '=', '(', ')', ',', 'dmu_at_0', '=', 'False', ',', 'return_se_dict', '=', 'False', ')', ':', 'chempot_range', '=', 'sorted... | Creates a dictionary where each entry is a key pointing to a
chemical potential range where the surface of that entry is stable.
Does so by enumerating through all possible solutions (intersect)
for surface energies of a specific facet.
Args:
chempot_range ([max_chempot, min... | ['Creates', 'a', 'dictionary', 'where', 'each', 'entry', 'is', 'a', 'key', 'pointing', 'to', 'a', 'chemical', 'potential', 'range', 'where', 'the', 'surface', 'of', 'that', 'entry', 'is', 'stable', '.', 'Does', 'so', 'by', 'enumerating', 'through', 'all', 'possible', 'solutions', '(', 'intersect', ')', 'for', 'surface'... | train | https://github.com/materialsproject/pymatgen/blob/4ca558cf72f8d5f8a1f21dfdfc0181a971c186da/pymatgen/analysis/surface_analysis.py#L661-L781 |
8,270 | pycontribs/pyrax | pyrax/object_storage.py | ContainerManager.delete_account_metadata | def delete_account_metadata(self, prefix=None):
"""
Removes all metadata matching the specified prefix from the account.
By default, the standard account metadata prefix ('X-Account-Meta-') is
prepended to the header name if it isn't present. For non-standard
headers, you must i... | python | def delete_account_metadata(self, prefix=None):
"""
Removes all metadata matching the specified prefix from the account.
By default, the standard account metadata prefix ('X-Account-Meta-') is
prepended to the header name if it isn't present. For non-standard
headers, you must i... | ['def', 'delete_account_metadata', '(', 'self', ',', 'prefix', '=', 'None', ')', ':', '# Add the metadata prefix, if needed.', 'if', 'prefix', 'is', 'None', ':', 'prefix', '=', 'ACCOUNT_META_PREFIX', 'curr_meta', '=', 'self', '.', 'get_account_metadata', '(', 'prefix', '=', 'prefix', ')', 'for', 'ckey', 'in', 'curr_met... | Removes all metadata matching the specified prefix from the account.
By default, the standard account metadata prefix ('X-Account-Meta-') is
prepended to the header name if it isn't present. For non-standard
headers, you must include a non-None prefix, such as an empty string. | ['Removes', 'all', 'metadata', 'matching', 'the', 'specified', 'prefix', 'from', 'the', 'account', '.'] | train | https://github.com/pycontribs/pyrax/blob/9ddfd5064b3a292d7337906f3b2d5dce95b50b99/pyrax/object_storage.py#L948-L965 |
8,271 | respeaker/respeaker_python_library | respeaker/usb_hid/hidapi_backend.py | HidApiUSB.write | def write(self, data):
"""
write data on the OUT endpoint associated to the HID interface
"""
for _ in range(64 - len(data)):
data.append(0)
#logging.debug("send: %s", data)
self.device.write(bytearray([0]) + data)
return | python | def write(self, data):
"""
write data on the OUT endpoint associated to the HID interface
"""
for _ in range(64 - len(data)):
data.append(0)
#logging.debug("send: %s", data)
self.device.write(bytearray([0]) + data)
return | ['def', 'write', '(', 'self', ',', 'data', ')', ':', 'for', '_', 'in', 'range', '(', '64', '-', 'len', '(', 'data', ')', ')', ':', 'data', '.', 'append', '(', '0', ')', '#logging.debug("send: %s", data)', 'self', '.', 'device', '.', 'write', '(', 'bytearray', '(', '[', '0', ']', ')', '+', 'data', ')', 'return'] | write data on the OUT endpoint associated to the HID interface | ['write', 'data', 'on', 'the', 'OUT', 'endpoint', 'associated', 'to', 'the', 'HID', 'interface'] | train | https://github.com/respeaker/respeaker_python_library/blob/905a5334ccdc2d474ad973caf6a23d05c65bbb25/respeaker/usb_hid/hidapi_backend.py#L98-L106 |
8,272 | michael-lazar/rtv | rtv/terminal.py | Terminal.strip_textpad | def strip_textpad(text):
"""
Attempt to intelligently strip excess whitespace from the output of a
curses textpad.
"""
if text is None:
return text
# Trivial case where the textbox is only one line long.
if '\n' not in text:
return text.r... | python | def strip_textpad(text):
"""
Attempt to intelligently strip excess whitespace from the output of a
curses textpad.
"""
if text is None:
return text
# Trivial case where the textbox is only one line long.
if '\n' not in text:
return text.r... | ['def', 'strip_textpad', '(', 'text', ')', ':', 'if', 'text', 'is', 'None', ':', 'return', 'text', '# Trivial case where the textbox is only one line long.', 'if', "'\\n'", 'not', 'in', 'text', ':', 'return', 'text', '.', 'rstrip', '(', ')', '# Allow one space at the end of the line. If there is more than one', '# spac... | Attempt to intelligently strip excess whitespace from the output of a
curses textpad. | ['Attempt', 'to', 'intelligently', 'strip', 'excess', 'whitespace', 'from', 'the', 'output', 'of', 'a', 'curses', 'textpad', '.'] | train | https://github.com/michael-lazar/rtv/blob/ccef2af042566ad384977028cf0bde01bc524dda/rtv/terminal.py#L878-L910 |
8,273 | fprimex/zdesk | zdesk/zdesk_api.py | ZendeskAPI.user_tickets_assigned | def user_tickets_assigned(self, user_id, external_id=None, **kwargs):
"https://developer.zendesk.com/rest_api/docs/core/tickets#allowed-for"
api_path = "/api/v2/users/{user_id}/tickets/assigned.json"
api_path = api_path.format(user_id=user_id)
api_query = {}
if "query" in kwargs.... | python | def user_tickets_assigned(self, user_id, external_id=None, **kwargs):
"https://developer.zendesk.com/rest_api/docs/core/tickets#allowed-for"
api_path = "/api/v2/users/{user_id}/tickets/assigned.json"
api_path = api_path.format(user_id=user_id)
api_query = {}
if "query" in kwargs.... | ['def', 'user_tickets_assigned', '(', 'self', ',', 'user_id', ',', 'external_id', '=', 'None', ',', '*', '*', 'kwargs', ')', ':', 'api_path', '=', '"/api/v2/users/{user_id}/tickets/assigned.json"', 'api_path', '=', 'api_path', '.', 'format', '(', 'user_id', '=', 'user_id', ')', 'api_query', '=', '{', '}', 'if', '"query... | https://developer.zendesk.com/rest_api/docs/core/tickets#allowed-for | ['https', ':', '//', 'developer', '.', 'zendesk', '.', 'com', '/', 'rest_api', '/', 'docs', '/', 'core', '/', 'tickets#allowed', '-', 'for'] | train | https://github.com/fprimex/zdesk/blob/851611c13b4d530e9df31390b3ec709baf0a0188/zdesk/zdesk_api.py#L4110-L4122 |
8,274 | cggh/scikit-allel | allel/compat.py | memoryview_safe | def memoryview_safe(x):
"""Make array safe to run in a Cython memoryview-based kernel. These
kernels typically break down with the error ``ValueError: buffer source
array is read-only`` when running in dask distributed.
See Also
--------
https://github.com/dask/distributed/issues/1978
https... | python | def memoryview_safe(x):
"""Make array safe to run in a Cython memoryview-based kernel. These
kernels typically break down with the error ``ValueError: buffer source
array is read-only`` when running in dask distributed.
See Also
--------
https://github.com/dask/distributed/issues/1978
https... | ['def', 'memoryview_safe', '(', 'x', ')', ':', 'if', 'not', 'x', '.', 'flags', '.', 'writeable', ':', 'if', 'not', 'x', '.', 'flags', '.', 'owndata', ':', 'x', '=', 'x', '.', 'copy', '(', 'order', '=', "'A'", ')', 'x', '.', 'setflags', '(', 'write', '=', 'True', ')', 'return', 'x'] | Make array safe to run in a Cython memoryview-based kernel. These
kernels typically break down with the error ``ValueError: buffer source
array is read-only`` when running in dask distributed.
See Also
--------
https://github.com/dask/distributed/issues/1978
https://github.com/cggh/scikit-allel... | ['Make', 'array', 'safe', 'to', 'run', 'in', 'a', 'Cython', 'memoryview', '-', 'based', 'kernel', '.', 'These', 'kernels', 'typically', 'break', 'down', 'with', 'the', 'error', 'ValueError', ':', 'buffer', 'source', 'array', 'is', 'read', '-', 'only', 'when', 'running', 'in', 'dask', 'distributed', '.'] | train | https://github.com/cggh/scikit-allel/blob/3c979a57a100240ba959dd13f98839349530f215/allel/compat.py#L51-L66 |
8,275 | ianmiell/shutit | shutit_pexpect.py | ShutItPexpectSession.whoami | def whoami(self,
note=None,
loglevel=logging.DEBUG):
"""Returns the current user by executing "whoami".
@param note: See send()
@return: the output of "whoami"
@rtype: string
"""
shutit = self.shutit
shutit.handle_note(note)
res = self.send_and_get_output(' command whoami',... | python | def whoami(self,
note=None,
loglevel=logging.DEBUG):
"""Returns the current user by executing "whoami".
@param note: See send()
@return: the output of "whoami"
@rtype: string
"""
shutit = self.shutit
shutit.handle_note(note)
res = self.send_and_get_output(' command whoami',... | ['def', 'whoami', '(', 'self', ',', 'note', '=', 'None', ',', 'loglevel', '=', 'logging', '.', 'DEBUG', ')', ':', 'shutit', '=', 'self', '.', 'shutit', 'shutit', '.', 'handle_note', '(', 'note', ')', 'res', '=', 'self', '.', 'send_and_get_output', '(', "' command whoami'", ',', 'echo', '=', 'False', ',', 'loglevel', '=... | Returns the current user by executing "whoami".
@param note: See send()
@return: the output of "whoami"
@rtype: string | ['Returns', 'the', 'current', 'user', 'by', 'executing', 'whoami', '.'] | train | https://github.com/ianmiell/shutit/blob/19cd64cdfb23515b106b40213dccff4101617076/shutit_pexpect.py#L671-L691 |
8,276 | santosjorge/cufflinks | cufflinks/colors.py | to_rgba | def to_rgba(color, alpha):
"""
Converts from hex|rgb to rgba
Parameters:
-----------
color : string
Color representation on hex or rgb
alpha : float
Value from 0 to 1.0 that represents the
alpha value.
Example:
... | python | def to_rgba(color, alpha):
"""
Converts from hex|rgb to rgba
Parameters:
-----------
color : string
Color representation on hex or rgb
alpha : float
Value from 0 to 1.0 that represents the
alpha value.
Example:
... | ['def', 'to_rgba', '(', 'color', ',', 'alpha', ')', ':', 'if', 'type', '(', 'color', ')', '==', 'tuple', ':', 'color', ',', 'alpha', '=', 'color', 'color', '=', 'color', '.', 'lower', '(', ')', 'if', "'rgba'", 'in', 'color', ':', 'cl', '=', 'list', '(', 'eval', '(', 'color', '.', 'replace', '(', "'rgba'", ',', "''", ')... | Converts from hex|rgb to rgba
Parameters:
-----------
color : string
Color representation on hex or rgb
alpha : float
Value from 0 to 1.0 that represents the
alpha value.
Example:
to_rgba('#E1E5ED',0.6)
... | ['Converts', 'from', 'hex|rgb', 'to', 'rgba'] | train | https://github.com/santosjorge/cufflinks/blob/ca1cbf93998dc793d0b1f8ac30fe1f2bd105f63a/cufflinks/colors.py#L23-L52 |
8,277 | linkedin/luminol | src/luminol/modules/time_series.py | TimeSeries.sum | def sum(self, default=None):
"""
Calculate the sum of all the values in the times series.
:param default: Value to return as a default should the calculation not be possible.
:return: Float representing the sum or `None`.
"""
return numpy.asscalar(numpy.sum(self.values))... | python | def sum(self, default=None):
"""
Calculate the sum of all the values in the times series.
:param default: Value to return as a default should the calculation not be possible.
:return: Float representing the sum or `None`.
"""
return numpy.asscalar(numpy.sum(self.values))... | ['def', 'sum', '(', 'self', ',', 'default', '=', 'None', ')', ':', 'return', 'numpy', '.', 'asscalar', '(', 'numpy', '.', 'sum', '(', 'self', '.', 'values', ')', ')', 'if', 'self', '.', 'values', 'else', 'default'] | Calculate the sum of all the values in the times series.
:param default: Value to return as a default should the calculation not be possible.
:return: Float representing the sum or `None`. | ['Calculate', 'the', 'sum', 'of', 'all', 'the', 'values', 'in', 'the', 'times', 'series', '.'] | train | https://github.com/linkedin/luminol/blob/42e4ab969b774ff98f902d064cb041556017f635/src/luminol/modules/time_series.py#L367-L374 |
8,278 | RJT1990/pyflux | pyflux/var/var.py | VAR._create_B | def _create_B(self,Y):
""" Creates OLS coefficient matrix
Parameters
----------
Y : np.array
The dependent variables Y
Returns
----------
The coefficient matrix B
"""
Z = self._create_Z(Y)
return np.dot(np.dot(Y,np.t... | python | def _create_B(self,Y):
""" Creates OLS coefficient matrix
Parameters
----------
Y : np.array
The dependent variables Y
Returns
----------
The coefficient matrix B
"""
Z = self._create_Z(Y)
return np.dot(np.dot(Y,np.t... | ['def', '_create_B', '(', 'self', ',', 'Y', ')', ':', 'Z', '=', 'self', '.', '_create_Z', '(', 'Y', ')', 'return', 'np', '.', 'dot', '(', 'np', '.', 'dot', '(', 'Y', ',', 'np', '.', 'transpose', '(', 'Z', ')', ')', ',', 'np', '.', 'linalg', '.', 'inv', '(', 'np', '.', 'dot', '(', 'Z', ',', 'np', '.', 'transpose', '(', ... | Creates OLS coefficient matrix
Parameters
----------
Y : np.array
The dependent variables Y
Returns
----------
The coefficient matrix B | ['Creates', 'OLS', 'coefficient', 'matrix'] | train | https://github.com/RJT1990/pyflux/blob/297f2afc2095acd97c12e827dd500e8ea5da0c0f/pyflux/var/var.py#L97-L111 |
8,279 | spyder-ide/spyder | spyder/plugins/ipythonconsole/widgets/client.py | ClientWidget.configure_shellwidget | def configure_shellwidget(self, give_focus=True):
"""Configure shellwidget after kernel is started"""
if give_focus:
self.get_control().setFocus()
# Set exit callback
self.shellwidget.set_exit_callback()
# To save history
self.shellwidget.executing.... | python | def configure_shellwidget(self, give_focus=True):
"""Configure shellwidget after kernel is started"""
if give_focus:
self.get_control().setFocus()
# Set exit callback
self.shellwidget.set_exit_callback()
# To save history
self.shellwidget.executing.... | ['def', 'configure_shellwidget', '(', 'self', ',', 'give_focus', '=', 'True', ')', ':', 'if', 'give_focus', ':', 'self', '.', 'get_control', '(', ')', '.', 'setFocus', '(', ')', '# Set exit callback\r', 'self', '.', 'shellwidget', '.', 'set_exit_callback', '(', ')', '# To save history\r', 'self', '.', 'shellwidget', '.... | Configure shellwidget after kernel is started | ['Configure', 'shellwidget', 'after', 'kernel', 'is', 'started'] | train | https://github.com/spyder-ide/spyder/blob/f76836ce1b924bcc4efd3f74f2960d26a4e528e0/spyder/plugins/ipythonconsole/widgets/client.py#L241-L292 |
8,280 | vburenin/xjpath | xjpath/xjpath.py | _full_sub_array | def _full_sub_array(data_obj, xj_path, create_dict_path):
"""Retrieves all array or dictionary elements for '*' JSON path marker.
:param dict|list data_obj: The current data object.
:param str xj_path: A json path.
:param bool create_dict_path create a dict path.
:return: tuple with two values: fir... | python | def _full_sub_array(data_obj, xj_path, create_dict_path):
"""Retrieves all array or dictionary elements for '*' JSON path marker.
:param dict|list data_obj: The current data object.
:param str xj_path: A json path.
:param bool create_dict_path create a dict path.
:return: tuple with two values: fir... | ['def', '_full_sub_array', '(', 'data_obj', ',', 'xj_path', ',', 'create_dict_path', ')', ':', 'if', 'isinstance', '(', 'data_obj', ',', 'list', ')', ':', 'if', 'xj_path', ':', 'res', '=', '[', ']', 'for', 'd', 'in', 'data_obj', ':', 'val', ',', 'exists', '=', 'path_lookup', '(', 'd', ',', 'xj_path', ',', 'create_dict_... | Retrieves all array or dictionary elements for '*' JSON path marker.
:param dict|list data_obj: The current data object.
:param str xj_path: A json path.
:param bool create_dict_path create a dict path.
:return: tuple with two values: first is a result and second
a boolean flag telling if ... | ['Retrieves', 'all', 'array', 'or', 'dictionary', 'elements', 'for', '*', 'JSON', 'path', 'marker', '.'] | train | https://github.com/vburenin/xjpath/blob/98a19fd6e6d0bcdc5ecbd3651ffa8915f06d7d44/xjpath/xjpath.py#L158-L189 |
8,281 | spotify/luigi | luigi/tools/range.py | infer_bulk_complete_from_fs | def infer_bulk_complete_from_fs(datetimes, datetime_to_task, datetime_to_re):
"""
Efficiently determines missing datetimes by filesystem listing.
The current implementation works for the common case of a task writing
output to a ``FileSystemTarget`` whose path is built using strftime with
format li... | python | def infer_bulk_complete_from_fs(datetimes, datetime_to_task, datetime_to_re):
"""
Efficiently determines missing datetimes by filesystem listing.
The current implementation works for the common case of a task writing
output to a ``FileSystemTarget`` whose path is built using strftime with
format li... | ['def', 'infer_bulk_complete_from_fs', '(', 'datetimes', ',', 'datetime_to_task', ',', 'datetime_to_re', ')', ':', 'filesystems_and_globs_by_location', '=', '_get_filesystems_and_globs', '(', 'datetime_to_task', ',', 'datetime_to_re', ')', 'paths_by_datetime', '=', '[', '[', 'o', '.', 'path', 'for', 'o', 'in', 'flatten... | Efficiently determines missing datetimes by filesystem listing.
The current implementation works for the common case of a task writing
output to a ``FileSystemTarget`` whose path is built using strftime with
format like '...%Y...%m...%d...%H...', without custom ``complete()`` or
``exists()``.
(Eve... | ['Efficiently', 'determines', 'missing', 'datetimes', 'by', 'filesystem', 'listing', '.'] | train | https://github.com/spotify/luigi/blob/c5eca1c3c3ee2a7eb612486192a0da146710a1e9/luigi/tools/range.py#L622-L647 |
8,282 | Karaage-Cluster/python-tldap | tldap/database/__init__.py | _db_to_python | def _db_to_python(db_data: dict, table: LdapObjectClass, dn: str) -> LdapObject:
""" Convert a DbDate object to a LdapObject. """
fields = table.get_fields()
python_data = table({
name: field.to_python(db_data[name])
for name, field in fields.items()
if field.db_field
})
pyt... | python | def _db_to_python(db_data: dict, table: LdapObjectClass, dn: str) -> LdapObject:
""" Convert a DbDate object to a LdapObject. """
fields = table.get_fields()
python_data = table({
name: field.to_python(db_data[name])
for name, field in fields.items()
if field.db_field
})
pyt... | ['def', '_db_to_python', '(', 'db_data', ':', 'dict', ',', 'table', ':', 'LdapObjectClass', ',', 'dn', ':', 'str', ')', '->', 'LdapObject', ':', 'fields', '=', 'table', '.', 'get_fields', '(', ')', 'python_data', '=', 'table', '(', '{', 'name', ':', 'field', '.', 'to_python', '(', 'db_data', '[', 'name', ']', ')', 'for... | Convert a DbDate object to a LdapObject. | ['Convert', 'a', 'DbDate', 'object', 'to', 'a', 'LdapObject', '.'] | train | https://github.com/Karaage-Cluster/python-tldap/blob/61f1af74a3648cb6491e7eeb1ee2eb395d67bf59/tldap/database/__init__.py#L392-L404 |
8,283 | bioinf-jku/FCD | build/lib/fcd/FCD.py | calculate_frechet_distance | def calculate_frechet_distance(mu1, sigma1, mu2, sigma2, eps=1e-6):
"""Numpy implementation of the Frechet Distance.
The Frechet distance between two multivariate Gaussians X_1 ~ N(mu_1, C_1)
and X_2 ~ N(mu_2, C_2) is
d^2 = ||mu_1 - mu_2||^2 + Tr(C_1 + C_2 - 2*sqrt(C_1*C_2)).
St... | python | def calculate_frechet_distance(mu1, sigma1, mu2, sigma2, eps=1e-6):
"""Numpy implementation of the Frechet Distance.
The Frechet distance between two multivariate Gaussians X_1 ~ N(mu_1, C_1)
and X_2 ~ N(mu_2, C_2) is
d^2 = ||mu_1 - mu_2||^2 + Tr(C_1 + C_2 - 2*sqrt(C_1*C_2)).
St... | ['def', 'calculate_frechet_distance', '(', 'mu1', ',', 'sigma1', ',', 'mu2', ',', 'sigma2', ',', 'eps', '=', '1e-6', ')', ':', 'mu1', '=', 'np', '.', 'atleast_1d', '(', 'mu1', ')', 'mu2', '=', 'np', '.', 'atleast_1d', '(', 'mu2', ')', 'sigma1', '=', 'np', '.', 'atleast_2d', '(', 'sigma1', ')', 'sigma2', '=', 'np', '.',... | Numpy implementation of the Frechet Distance.
The Frechet distance between two multivariate Gaussians X_1 ~ N(mu_1, C_1)
and X_2 ~ N(mu_2, C_2) is
d^2 = ||mu_1 - mu_2||^2 + Tr(C_1 + C_2 - 2*sqrt(C_1*C_2)).
Stable version by Dougal J. Sutherland.
Params:
-- mu1: The mean ... | ['Numpy', 'implementation', 'of', 'the', 'Frechet', 'Distance', '.', 'The', 'Frechet', 'distance', 'between', 'two', 'multivariate', 'Gaussians', 'X_1', '~', 'N', '(', 'mu_1', 'C_1', ')', 'and', 'X_2', '~', 'N', '(', 'mu_2', 'C_2', ')', 'is', 'd^2', '=', '||mu_1', '-', 'mu_2||^2', '+', 'Tr', '(', 'C_1', '+', 'C_2', '-'... | train | https://github.com/bioinf-jku/FCD/blob/fe542b16d72a2d0899989374e1a86cc930d891e1/build/lib/fcd/FCD.py#L33-L85 |
8,284 | bcbio/bcbio-nextgen | bcbio/distributed/objectstore.py | AzureBlob.download | def download(cls, filename, input_dir, dl_dir=None):
"""Download the resource from the storage."""
file_info = cls.parse_remote(filename)
if not dl_dir:
dl_dir = os.path.join(input_dir, file_info.container,
os.path.dirname(file_info.blob))
... | python | def download(cls, filename, input_dir, dl_dir=None):
"""Download the resource from the storage."""
file_info = cls.parse_remote(filename)
if not dl_dir:
dl_dir = os.path.join(input_dir, file_info.container,
os.path.dirname(file_info.blob))
... | ['def', 'download', '(', 'cls', ',', 'filename', ',', 'input_dir', ',', 'dl_dir', '=', 'None', ')', ':', 'file_info', '=', 'cls', '.', 'parse_remote', '(', 'filename', ')', 'if', 'not', 'dl_dir', ':', 'dl_dir', '=', 'os', '.', 'path', '.', 'join', '(', 'input_dir', ',', 'file_info', '.', 'container', ',', 'os', '.', 'p... | Download the resource from the storage. | ['Download', 'the', 'resource', 'from', 'the', 'storage', '.'] | train | https://github.com/bcbio/bcbio-nextgen/blob/6a9348c0054ccd5baffd22f1bb7d0422f6978b20/bcbio/distributed/objectstore.py#L497-L514 |
8,285 | bitcraze/crazyflie-lib-python | cflib/crazyflie/__init__.py | Crazyflie._start_connection_setup | def _start_connection_setup(self):
"""Start the connection setup by refreshing the TOCs"""
logger.info('We are connected[%s], request connection setup',
self.link_uri)
self.platform.fetch_platform_informations(self._platform_info_fetched) | python | def _start_connection_setup(self):
"""Start the connection setup by refreshing the TOCs"""
logger.info('We are connected[%s], request connection setup',
self.link_uri)
self.platform.fetch_platform_informations(self._platform_info_fetched) | ['def', '_start_connection_setup', '(', 'self', ')', ':', 'logger', '.', 'info', '(', "'We are connected[%s], request connection setup'", ',', 'self', '.', 'link_uri', ')', 'self', '.', 'platform', '.', 'fetch_platform_informations', '(', 'self', '.', '_platform_info_fetched', ')'] | Start the connection setup by refreshing the TOCs | ['Start', 'the', 'connection', 'setup', 'by', 'refreshing', 'the', 'TOCs'] | train | https://github.com/bitcraze/crazyflie-lib-python/blob/f6ebb4eb315bbe6e02db518936ac17fb615b2af8/cflib/crazyflie/__init__.py#L156-L160 |
8,286 | fitodic/centerline | centerline/main.py | Centerline._create_centerline | def _create_centerline(self):
"""
Calculate the centerline of a polygon.
Densifies the border of a polygon which is then represented by a Numpy
array of points necessary for creating the Voronoi diagram. Once the
diagram is created, the ridges located within the polygon are
... | python | def _create_centerline(self):
"""
Calculate the centerline of a polygon.
Densifies the border of a polygon which is then represented by a Numpy
array of points necessary for creating the Voronoi diagram. Once the
diagram is created, the ridges located within the polygon are
... | ['def', '_create_centerline', '(', 'self', ')', ':', 'border', '=', 'array', '(', 'self', '.', '__densify_border', '(', ')', ')', 'vor', '=', 'Voronoi', '(', 'border', ')', 'vertex', '=', 'vor', '.', 'vertices', 'lst_lines', '=', '[', ']', 'for', 'j', ',', 'ridge', 'in', 'enumerate', '(', 'vor', '.', 'ridge_vertices', ... | Calculate the centerline of a polygon.
Densifies the border of a polygon which is then represented by a Numpy
array of points necessary for creating the Voronoi diagram. Once the
diagram is created, the ridges located within the polygon are
joined and returned.
Returns:
... | ['Calculate', 'the', 'centerline', 'of', 'a', 'polygon', '.'] | train | https://github.com/fitodic/centerline/blob/f27e7b1ecb77bd4da40093ab44754cbd3ec9f58b/centerline/main.py#L62-L99 |
8,287 | AndresMWeber/Nomenclate | nomenclate/core/tools.py | flatten | def flatten(it):
""" Flattens any iterable
From:
http://stackoverflow.com/questions/11503065/python-function-to-flatten-generator-containing-another-generator
:param it: Iterator, iterator to flatten
:return: Generator, A generator of the flattened values
"""
for x in it:
if... | python | def flatten(it):
""" Flattens any iterable
From:
http://stackoverflow.com/questions/11503065/python-function-to-flatten-generator-containing-another-generator
:param it: Iterator, iterator to flatten
:return: Generator, A generator of the flattened values
"""
for x in it:
if... | ['def', 'flatten', '(', 'it', ')', ':', 'for', 'x', 'in', 'it', ':', 'if', 'isinstance', '(', 'x', ',', 'collections', '.', 'Iterable', ')', 'and', 'not', 'isinstance', '(', 'x', ',', 'str', ')', ':', 'for', 'y', 'in', 'flatten', '(', 'x', ')', ':', 'yield', 'y', 'else', ':', 'yield', 'x'] | Flattens any iterable
From:
http://stackoverflow.com/questions/11503065/python-function-to-flatten-generator-containing-another-generator
:param it: Iterator, iterator to flatten
:return: Generator, A generator of the flattened values | ['Flattens', 'any', 'iterable', 'From', ':', 'http', ':', '//', 'stackoverflow', '.', 'com', '/', 'questions', '/', '11503065', '/', 'python', '-', 'function', '-', 'to', '-', 'flatten', '-', 'generator', '-', 'containing', '-', 'another', '-', 'generator'] | train | https://github.com/AndresMWeber/Nomenclate/blob/e6d6fc28beac042bad588e56fbe77531d2de6b6f/nomenclate/core/tools.py#L124-L137 |
8,288 | saltstack/salt | salt/cli/cp.py | SaltCP._recurse | def _recurse(self, path):
'''
Get a list of all specified files
'''
files = {}
empty_dirs = []
try:
sub_paths = os.listdir(path)
except OSError as exc:
if exc.errno == errno.ENOENT:
# Path does not exist
sys.... | python | def _recurse(self, path):
'''
Get a list of all specified files
'''
files = {}
empty_dirs = []
try:
sub_paths = os.listdir(path)
except OSError as exc:
if exc.errno == errno.ENOENT:
# Path does not exist
sys.... | ['def', '_recurse', '(', 'self', ',', 'path', ')', ':', 'files', '=', '{', '}', 'empty_dirs', '=', '[', ']', 'try', ':', 'sub_paths', '=', 'os', '.', 'listdir', '(', 'path', ')', 'except', 'OSError', 'as', 'exc', ':', 'if', 'exc', '.', 'errno', '==', 'errno', '.', 'ENOENT', ':', '# Path does not exist', 'sys', '.', 'st... | Get a list of all specified files | ['Get', 'a', 'list', 'of', 'all', 'specified', 'files'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/cli/cp.py#L71-L95 |
8,289 | pgjones/quart | quart/app.py | Quart.after_request | def after_request(self, func: Callable, name: AppOrBlueprintKey=None) -> Callable:
"""Add an after request function.
This is designed to be used as a decorator. An example usage,
.. code-block:: python
@app.after_request
def func(response):
return respo... | python | def after_request(self, func: Callable, name: AppOrBlueprintKey=None) -> Callable:
"""Add an after request function.
This is designed to be used as a decorator. An example usage,
.. code-block:: python
@app.after_request
def func(response):
return respo... | ['def', 'after_request', '(', 'self', ',', 'func', ':', 'Callable', ',', 'name', ':', 'AppOrBlueprintKey', '=', 'None', ')', '->', 'Callable', ':', 'handler', '=', 'ensure_coroutine', '(', 'func', ')', 'self', '.', 'after_request_funcs', '[', 'name', ']', '.', 'append', '(', 'handler', ')', 'return', 'func'] | Add an after request function.
This is designed to be used as a decorator. An example usage,
.. code-block:: python
@app.after_request
def func(response):
return response
Arguments:
func: The after request function itself.
name:... | ['Add', 'an', 'after', 'request', 'function', '.'] | train | https://github.com/pgjones/quart/blob/7cb2d3bd98e8746025764f2b933abc12041fa175/quart/app.py#L1085-L1102 |
8,290 | tagcubeio/tagcube-cli | tagcube_cli/utils.py | _parse_config_file_impl | def _parse_config_file_impl(filename):
"""
Format for the file is:
credentials:
email: ...
api_token: ...
:param filename: The filename to parse
:return: A tuple with:
- email
- api_token
"""
api_key = None
email = ... | python | def _parse_config_file_impl(filename):
"""
Format for the file is:
credentials:
email: ...
api_token: ...
:param filename: The filename to parse
:return: A tuple with:
- email
- api_token
"""
api_key = None
email = ... | ['def', '_parse_config_file_impl', '(', 'filename', ')', ':', 'api_key', '=', 'None', 'email', '=', 'None', 'try', ':', 'doc', '=', 'yaml', '.', 'load', '(', 'file', '(', 'filename', ')', '.', 'read', '(', ')', ')', 'email', '=', 'doc', '[', "'credentials'", ']', '[', "'email'", ']', 'api_key', '=', 'doc', '[', "'crede... | Format for the file is:
credentials:
email: ...
api_token: ...
:param filename: The filename to parse
:return: A tuple with:
- email
- api_token | ['Format', 'for', 'the', 'file', 'is', ':', 'credentials', ':', 'email', ':', '...', 'api_token', ':', '...', ':', 'param', 'filename', ':', 'The', 'filename', 'to', 'parse', ':', 'return', ':', 'A', 'tuple', 'with', ':', '-', 'email', '-', 'api_token'] | train | https://github.com/tagcubeio/tagcube-cli/blob/709e4b0b11331a4d2791dc79107e5081518d75bf/tagcube_cli/utils.py#L76-L117 |
8,291 | tanghaibao/jcvi | jcvi/formats/fastq.py | shuffle | def shuffle(args):
"""
%prog shuffle p1.fastq p2.fastq
Shuffle pairs into interleaved format.
"""
p = OptionParser(shuffle.__doc__)
p.set_tag()
opts, args = p.parse_args(args)
if len(args) != 2:
sys.exit(not p.print_help())
p1, p2 = args
pairsfastq = pairspf((p1, p2)) ... | python | def shuffle(args):
"""
%prog shuffle p1.fastq p2.fastq
Shuffle pairs into interleaved format.
"""
p = OptionParser(shuffle.__doc__)
p.set_tag()
opts, args = p.parse_args(args)
if len(args) != 2:
sys.exit(not p.print_help())
p1, p2 = args
pairsfastq = pairspf((p1, p2)) ... | ['def', 'shuffle', '(', 'args', ')', ':', 'p', '=', 'OptionParser', '(', 'shuffle', '.', '__doc__', ')', 'p', '.', 'set_tag', '(', ')', 'opts', ',', 'args', '=', 'p', '.', 'parse_args', '(', 'args', ')', 'if', 'len', '(', 'args', ')', '!=', '2', ':', 'sys', '.', 'exit', '(', 'not', 'p', '.', 'print_help', '(', ')', ')'... | %prog shuffle p1.fastq p2.fastq
Shuffle pairs into interleaved format. | ['%prog', 'shuffle', 'p1', '.', 'fastq', 'p2', '.', 'fastq'] | train | https://github.com/tanghaibao/jcvi/blob/d2e31a77b6ade7f41f3b321febc2b4744d1cdeca/jcvi/formats/fastq.py#L509-L551 |
8,292 | quantopian/zipline | zipline/finance/ledger.py | Ledger.process_transaction | def process_transaction(self, transaction):
"""Add a transaction to ledger, updating the current state as needed.
Parameters
----------
transaction : zp.Transaction
The transaction to execute.
"""
asset = transaction.asset
if isinstance(asset, Future)... | python | def process_transaction(self, transaction):
"""Add a transaction to ledger, updating the current state as needed.
Parameters
----------
transaction : zp.Transaction
The transaction to execute.
"""
asset = transaction.asset
if isinstance(asset, Future)... | ['def', 'process_transaction', '(', 'self', ',', 'transaction', ')', ':', 'asset', '=', 'transaction', '.', 'asset', 'if', 'isinstance', '(', 'asset', ',', 'Future', ')', ':', 'try', ':', 'old_price', '=', 'self', '.', '_payout_last_sale_prices', '[', 'asset', ']', 'except', 'KeyError', ':', 'self', '.', '_payout_last_... | Add a transaction to ledger, updating the current state as needed.
Parameters
----------
transaction : zp.Transaction
The transaction to execute. | ['Add', 'a', 'transaction', 'to', 'ledger', 'updating', 'the', 'current', 'state', 'as', 'needed', '.'] | train | https://github.com/quantopian/zipline/blob/77ad15e6dc4c1cbcdc133653bac8a63fc704f7fe/zipline/finance/ledger.py#L479-L523 |
8,293 | twisted/axiom | axiom/dependency.py | installedOn | def installedOn(self):
"""
If this item is installed on another item, return the install
target. Otherwise return None.
"""
try:
return self.store.findUnique(_DependencyConnector,
_DependencyConnector.installee == self
... | python | def installedOn(self):
"""
If this item is installed on another item, return the install
target. Otherwise return None.
"""
try:
return self.store.findUnique(_DependencyConnector,
_DependencyConnector.installee == self
... | ['def', 'installedOn', '(', 'self', ')', ':', 'try', ':', 'return', 'self', '.', 'store', '.', 'findUnique', '(', '_DependencyConnector', ',', '_DependencyConnector', '.', 'installee', '==', 'self', ')', '.', 'target', 'except', 'ItemNotFound', ':', 'return', 'None'] | If this item is installed on another item, return the install
target. Otherwise return None. | ['If', 'this', 'item', 'is', 'installed', 'on', 'another', 'item', 'return', 'the', 'install', 'target', '.', 'Otherwise', 'return', 'None', '.'] | train | https://github.com/twisted/axiom/blob/7de70bc8fe1bb81f9c2339fba8daec9eb2e92b68/axiom/dependency.py#L174-L184 |
8,294 | dnanexus/dx-toolkit | src/python/dxpy/api.py | global_workflow_run | def global_workflow_run(name_or_id, alias=None, input_params={}, always_retry=True, **kwargs):
"""
Invokes the /globalworkflow-xxxx/run API method.
For more info, see: https://wiki.dnanexus.com/API-Specification-v1.0.0/Global-Workflows#API-method:-/globalworkflow-xxxx%5B/yyyy%5D/run
"""
input_param... | python | def global_workflow_run(name_or_id, alias=None, input_params={}, always_retry=True, **kwargs):
"""
Invokes the /globalworkflow-xxxx/run API method.
For more info, see: https://wiki.dnanexus.com/API-Specification-v1.0.0/Global-Workflows#API-method:-/globalworkflow-xxxx%5B/yyyy%5D/run
"""
input_param... | ['def', 'global_workflow_run', '(', 'name_or_id', ',', 'alias', '=', 'None', ',', 'input_params', '=', '{', '}', ',', 'always_retry', '=', 'True', ',', '*', '*', 'kwargs', ')', ':', 'input_params_cp', '=', 'Nonce', '.', 'update_nonce', '(', 'input_params', ')', 'fully_qualified_version', '=', 'name_or_id', '+', '(', '(... | Invokes the /globalworkflow-xxxx/run API method.
For more info, see: https://wiki.dnanexus.com/API-Specification-v1.0.0/Global-Workflows#API-method:-/globalworkflow-xxxx%5B/yyyy%5D/run | ['Invokes', 'the', '/', 'globalworkflow', '-', 'xxxx', '/', 'run', 'API', 'method', '.'] | train | https://github.com/dnanexus/dx-toolkit/blob/74befb53ad90fcf902d8983ae6d74580f402d619/src/python/dxpy/api.py#L762-L770 |
8,295 | jldbc/pybaseball | pybaseball/statcast_batter.py | statcast_batter | def statcast_batter(start_dt=None, end_dt=None, player_id=None):
"""
Pulls statcast pitch-level data from Baseball Savant for a given batter.
ARGUMENTS
start_dt : YYYY-MM-DD : the first date for which you want a player's statcast data
end_dt : YYYY-MM-DD : the final date for which you want data
... | python | def statcast_batter(start_dt=None, end_dt=None, player_id=None):
"""
Pulls statcast pitch-level data from Baseball Savant for a given batter.
ARGUMENTS
start_dt : YYYY-MM-DD : the first date for which you want a player's statcast data
end_dt : YYYY-MM-DD : the final date for which you want data
... | ['def', 'statcast_batter', '(', 'start_dt', '=', 'None', ',', 'end_dt', '=', 'None', ',', 'player_id', '=', 'None', ')', ':', 'start_dt', ',', 'end_dt', ',', 'player_id', '=', 'sanitize_input', '(', 'start_dt', ',', 'end_dt', ',', 'player_id', ')', '# inputs are valid if either both or zero dates are supplied. Not vali... | Pulls statcast pitch-level data from Baseball Savant for a given batter.
ARGUMENTS
start_dt : YYYY-MM-DD : the first date for which you want a player's statcast data
end_dt : YYYY-MM-DD : the final date for which you want data
player_id : INT : the player's MLBAM ID. Find this by calling pybaseball.pla... | ['Pulls', 'statcast', 'pitch', '-', 'level', 'data', 'from', 'Baseball', 'Savant', 'for', 'a', 'given', 'batter', '.'] | train | https://github.com/jldbc/pybaseball/blob/085ea26bfd1b5f5926d79d4fac985c88278115f2/pybaseball/statcast_batter.py#L4-L18 |
8,296 | markovmodel/PyEMMA | devtools/ci/jenkins/update_versions_json.py | main | def main(argv=None):
'''Command line options.'''
if argv is None:
argv = sys.argv
else:
sys.argv.extend(argv)
parser = ArgumentParser()
parser.add_argument('-u', '--url', dest='url', required=True, help="base url (has to contain versions json)")
parser.add_argument('-o', '--out... | python | def main(argv=None):
'''Command line options.'''
if argv is None:
argv = sys.argv
else:
sys.argv.extend(argv)
parser = ArgumentParser()
parser.add_argument('-u', '--url', dest='url', required=True, help="base url (has to contain versions json)")
parser.add_argument('-o', '--out... | ['def', 'main', '(', 'argv', '=', 'None', ')', ':', 'if', 'argv', 'is', 'None', ':', 'argv', '=', 'sys', '.', 'argv', 'else', ':', 'sys', '.', 'argv', '.', 'extend', '(', 'argv', ')', 'parser', '=', 'ArgumentParser', '(', ')', 'parser', '.', 'add_argument', '(', "'-u'", ',', "'--url'", ',', 'dest', '=', "'url'", ',', '... | Command line options. | ['Command', 'line', 'options', '.'] | train | https://github.com/markovmodel/PyEMMA/blob/5c3124398217de05ba5ce9c8fb01519222481ab8/devtools/ci/jenkins/update_versions_json.py#L33-L82 |
8,297 | gwastro/pycbc | pycbc/types/frequencyseries.py | FrequencySeries.save | def save(self, path, group=None, ifo='P1'):
"""
Save frequency series to a Numpy .npy, hdf, or text file. The first column
contains the sample frequencies, the second contains the values.
In the case of a complex frequency series saved as text, the imaginary
part is written as a ... | python | def save(self, path, group=None, ifo='P1'):
"""
Save frequency series to a Numpy .npy, hdf, or text file. The first column
contains the sample frequencies, the second contains the values.
In the case of a complex frequency series saved as text, the imaginary
part is written as a ... | ['def', 'save', '(', 'self', ',', 'path', ',', 'group', '=', 'None', ',', 'ifo', '=', "'P1'", ')', ':', 'ext', '=', '_os', '.', 'path', '.', 'splitext', '(', 'path', ')', '[', '1', ']', 'if', 'ext', '==', "'.npy'", ':', 'output', '=', '_numpy', '.', 'vstack', '(', '(', 'self', '.', 'sample_frequencies', '.', 'numpy', '... | Save frequency series to a Numpy .npy, hdf, or text file. The first column
contains the sample frequencies, the second contains the values.
In the case of a complex frequency series saved as text, the imaginary
part is written as a third column. When using hdf format, the data is stored
... | ['Save', 'frequency', 'series', 'to', 'a', 'Numpy', '.', 'npy', 'hdf', 'or', 'text', 'file', '.', 'The', 'first', 'column', 'contains', 'the', 'sample', 'frequencies', 'the', 'second', 'contains', 'the', 'values', '.', 'In', 'the', 'case', 'of', 'a', 'complex', 'frequency', 'series', 'saved', 'as', 'text', 'the', 'imag... | train | https://github.com/gwastro/pycbc/blob/7a64cdd104d263f1b6ea0b01e6841837d05a4cb3/pycbc/types/frequencyseries.py#L366-L429 |
8,298 | spacetelescope/drizzlepac | drizzlepac/wcs_functions.py | WCSMap.forward | def forward(self,pixx,pixy):
""" Transform the input pixx,pixy positions in the input frame
to pixel positions in the output frame.
This method gets passed to the drizzle algorithm.
"""
# This matches WTRAXY results to better than 1e-4 pixels.
skyx,skyy = self.in... | python | def forward(self,pixx,pixy):
""" Transform the input pixx,pixy positions in the input frame
to pixel positions in the output frame.
This method gets passed to the drizzle algorithm.
"""
# This matches WTRAXY results to better than 1e-4 pixels.
skyx,skyy = self.in... | ['def', 'forward', '(', 'self', ',', 'pixx', ',', 'pixy', ')', ':', '# This matches WTRAXY results to better than 1e-4 pixels.', 'skyx', ',', 'skyy', '=', 'self', '.', 'input', '.', 'all_pix2world', '(', 'pixx', ',', 'pixy', ',', 'self', '.', 'origin', ')', 'result', '=', 'self', '.', 'output', '.', 'wcs_world2pix', '(... | Transform the input pixx,pixy positions in the input frame
to pixel positions in the output frame.
This method gets passed to the drizzle algorithm. | ['Transform', 'the', 'input', 'pixx', 'pixy', 'positions', 'in', 'the', 'input', 'frame', 'to', 'pixel', 'positions', 'in', 'the', 'output', 'frame', '.'] | train | https://github.com/spacetelescope/drizzlepac/blob/15bec3c929a6a869d9e71b9398ced43ede0620f1/drizzlepac/wcs_functions.py#L53-L62 |
8,299 | browniebroke/deezer-python | deezer/client.py | Client.get_object | def get_object(
self, object_t, object_id=None, relation=None, parent=None, **kwargs
):
"""
Actually query the Deezer API to retrieve the object
:returns: json dictionary
"""
url = self.object_url(object_t, object_id, relation, **kwargs)
response = self.sessi... | python | def get_object(
self, object_t, object_id=None, relation=None, parent=None, **kwargs
):
"""
Actually query the Deezer API to retrieve the object
:returns: json dictionary
"""
url = self.object_url(object_t, object_id, relation, **kwargs)
response = self.sessi... | ['def', 'get_object', '(', 'self', ',', 'object_t', ',', 'object_id', '=', 'None', ',', 'relation', '=', 'None', ',', 'parent', '=', 'None', ',', '*', '*', 'kwargs', ')', ':', 'url', '=', 'self', '.', 'object_url', '(', 'object_t', ',', 'object_id', ',', 'relation', ',', '*', '*', 'kwargs', ')', 'response', '=', 'self'... | Actually query the Deezer API to retrieve the object
:returns: json dictionary | ['Actually', 'query', 'the', 'Deezer', 'API', 'to', 'retrieve', 'the', 'object'] | train | https://github.com/browniebroke/deezer-python/blob/fb869c3617045b22e7124e4b783ec1a68d283ac3/deezer/client.py#L128-L138 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.