Unnamed: 0 int64 0 10k | repository_name stringlengths 7 54 | func_path_in_repository stringlengths 5 223 | func_name stringlengths 1 134 | whole_func_string stringlengths 100 30.3k | language stringclasses 1
value | func_code_string stringlengths 100 30.3k | func_code_tokens stringlengths 138 33.2k | func_documentation_string stringlengths 1 15k | func_documentation_tokens stringlengths 5 5.14k | split_name stringclasses 1
value | func_code_url stringlengths 91 315 |
|---|---|---|---|---|---|---|---|---|---|---|---|
4,500 | mikicz/arca | arca/backend/current_environment.py | CurrentEnvironmentBackend.get_or_create_environment | def get_or_create_environment(self, repo: str, branch: str, git_repo: Repo, repo_path: Path) -> str:
""" Returns the path to the current Python executable.
"""
return sys.executable | python | def get_or_create_environment(self, repo: str, branch: str, git_repo: Repo, repo_path: Path) -> str:
""" Returns the path to the current Python executable.
"""
return sys.executable | ['def', 'get_or_create_environment', '(', 'self', ',', 'repo', ':', 'str', ',', 'branch', ':', 'str', ',', 'git_repo', ':', 'Repo', ',', 'repo_path', ':', 'Path', ')', '->', 'str', ':', 'return', 'sys', '.', 'executable'] | Returns the path to the current Python executable. | ['Returns', 'the', 'path', 'to', 'the', 'current', 'Python', 'executable', '.'] | train | https://github.com/mikicz/arca/blob/e67fdc00be473ecf8ec16d024e1a3f2c47ca882c/arca/backend/current_environment.py#L15-L18 |
4,501 | ktbyers/netmiko | netmiko/flexvnf/flexvnf_ssh.py | FlexvnfSSH.commit | def commit(
self,
confirm=False,
confirm_delay=None,
check=False,
comment="",
and_quit=False,
delay_factor=1,
):
"""
Commit the candidate configuration.
Commit the entered configuration. Raise an error and return the failure
if... | python | def commit(
self,
confirm=False,
confirm_delay=None,
check=False,
comment="",
and_quit=False,
delay_factor=1,
):
"""
Commit the candidate configuration.
Commit the entered configuration. Raise an error and return the failure
if... | ['def', 'commit', '(', 'self', ',', 'confirm', '=', 'False', ',', 'confirm_delay', '=', 'None', ',', 'check', '=', 'False', ',', 'comment', '=', '""', ',', 'and_quit', '=', 'False', ',', 'delay_factor', '=', '1', ',', ')', ':', 'delay_factor', '=', 'self', '.', 'select_delay_factor', '(', 'delay_factor', ')', 'if', 'ch... | Commit the candidate configuration.
Commit the entered configuration. Raise an error and return the failure
if the commit fails.
Automatically enters configuration mode
default:
command_string = commit
check and (confirm or confirm_dely or comment):
Exc... | ['Commit', 'the', 'candidate', 'configuration', '.'] | train | https://github.com/ktbyers/netmiko/blob/54e6116c0b4664de2123081937e0a9a27bdfdfea/netmiko/flexvnf/flexvnf_ssh.py#L82-L171 |
4,502 | marshmallow-code/marshmallow | src/marshmallow/schema.py | BaseSchema._do_load | def _do_load(
self, data, many=None, partial=None, unknown=None,
postprocess=True,
):
"""Deserialize `data`, returning the deserialized result.
:param data: The data to deserialize.
:param bool many: Whether to deserialize `data` as a collection. If `None`, the
v... | python | def _do_load(
self, data, many=None, partial=None, unknown=None,
postprocess=True,
):
"""Deserialize `data`, returning the deserialized result.
:param data: The data to deserialize.
:param bool many: Whether to deserialize `data` as a collection. If `None`, the
v... | ['def', '_do_load', '(', 'self', ',', 'data', ',', 'many', '=', 'None', ',', 'partial', '=', 'None', ',', 'unknown', '=', 'None', ',', 'postprocess', '=', 'True', ',', ')', ':', 'error_store', '=', 'ErrorStore', '(', ')', 'errors', '=', '{', '}', 'many', '=', 'self', '.', 'many', 'if', 'many', 'is', 'None', 'else', 'bo... | Deserialize `data`, returning the deserialized result.
:param data: The data to deserialize.
:param bool many: Whether to deserialize `data` as a collection. If `None`, the
value for `self.many` is used.
:param bool|tuple partial: Whether to validate required fields. If its value is... | ['Deserialize', 'data', 'returning', 'the', 'deserialized', 'result', '.'] | train | https://github.com/marshmallow-code/marshmallow/blob/a6b6c4151f1fbf16f3774d4052ca2bddf6903750/src/marshmallow/schema.py#L776-L870 |
4,503 | CodyKochmann/graphdb | graphdb/RamGraphDB.py | VList._where | def _where(self, filter_fn):
''' use this to filter VLists, simply provide a filter function to filter the current found objects '''
assert callable(filter_fn), 'filter_fn needs to be callable'
return VList(i for i in self if filter_fn(i())) | python | def _where(self, filter_fn):
''' use this to filter VLists, simply provide a filter function to filter the current found objects '''
assert callable(filter_fn), 'filter_fn needs to be callable'
return VList(i for i in self if filter_fn(i())) | ['def', '_where', '(', 'self', ',', 'filter_fn', ')', ':', 'assert', 'callable', '(', 'filter_fn', ')', ',', "'filter_fn needs to be callable'", 'return', 'VList', '(', 'i', 'for', 'i', 'in', 'self', 'if', 'filter_fn', '(', 'i', '(', ')', ')', ')'] | use this to filter VLists, simply provide a filter function to filter the current found objects | ['use', 'this', 'to', 'filter', 'VLists', 'simply', 'provide', 'a', 'filter', 'function', 'to', 'filter', 'the', 'current', 'found', 'objects'] | train | https://github.com/CodyKochmann/graphdb/blob/8c18830db4beda30204f5fd4450bc96eb39b0feb/graphdb/RamGraphDB.py#L355-L358 |
4,504 | Microsoft/LightGBM | python-package/lightgbm/basic.py | Booster.dump_model | def dump_model(self, num_iteration=None, start_iteration=0):
"""Dump Booster to JSON format.
Parameters
----------
num_iteration : int or None, optional (default=None)
Index of the iteration that should be dumped.
If None, if the best iteration exists, it is dump... | python | def dump_model(self, num_iteration=None, start_iteration=0):
"""Dump Booster to JSON format.
Parameters
----------
num_iteration : int or None, optional (default=None)
Index of the iteration that should be dumped.
If None, if the best iteration exists, it is dump... | ['def', 'dump_model', '(', 'self', ',', 'num_iteration', '=', 'None', ',', 'start_iteration', '=', '0', ')', ':', 'if', 'num_iteration', 'is', 'None', ':', 'num_iteration', '=', 'self', '.', 'best_iteration', 'buffer_len', '=', '1', '<<', '20', 'tmp_out_len', '=', 'ctypes', '.', 'c_int64', '(', '0', ')', 'string_buffer... | Dump Booster to JSON format.
Parameters
----------
num_iteration : int or None, optional (default=None)
Index of the iteration that should be dumped.
If None, if the best iteration exists, it is dumped; otherwise, all iterations are dumped.
If <= 0, all itera... | ['Dump', 'Booster', 'to', 'JSON', 'format', '.'] | train | https://github.com/Microsoft/LightGBM/blob/8d2ec69f4f685b0ab1c4624d59ee2d3287bb3147/python-package/lightgbm/basic.py#L2194-L2239 |
4,505 | PatrikValkovic/grammpy | grammpy/transforms/Manipulations.py | Manipulations.replace | def replace(oldEl, newEl):
# type: (Union[Rule, _RuleConnectable], Union[Rule, _RuleConnectable]) -> Union[Rule, _RuleConnectable]
"""
Replace element in the parsed tree. Can be nonterminal, terminal or rule.
:param oldEl: Element already in the tree.
:param newEl: Element to rep... | python | def replace(oldEl, newEl):
# type: (Union[Rule, _RuleConnectable], Union[Rule, _RuleConnectable]) -> Union[Rule, _RuleConnectable]
"""
Replace element in the parsed tree. Can be nonterminal, terminal or rule.
:param oldEl: Element already in the tree.
:param newEl: Element to rep... | ['def', 'replace', '(', 'oldEl', ',', 'newEl', ')', ':', '# type: (Union[Rule, _RuleConnectable], Union[Rule, _RuleConnectable]) -> Union[Rule, _RuleConnectable]', 'if', 'isinstance', '(', 'oldEl', ',', 'Rule', ')', ':', 'return', 'Manipulations', '.', 'replaceRule', '(', 'oldEl', ',', 'newEl', ')', 'if', 'isinstance',... | Replace element in the parsed tree. Can be nonterminal, terminal or rule.
:param oldEl: Element already in the tree.
:param newEl: Element to replace with.
:return: New element attached to the tree. | ['Replace', 'element', 'in', 'the', 'parsed', 'tree', '.', 'Can', 'be', 'nonterminal', 'terminal', 'or', 'rule', '.', ':', 'param', 'oldEl', ':', 'Element', 'already', 'in', 'the', 'tree', '.', ':', 'param', 'newEl', ':', 'Element', 'to', 'replace', 'with', '.', ':', 'return', ':', 'New', 'element', 'attached', 'to', '... | train | https://github.com/PatrikValkovic/grammpy/blob/879ce0ef794ac2823acc19314fcd7a8aba53e50f/grammpy/transforms/Manipulations.py#L58-L69 |
4,506 | nicferrier/md | src/mdlib/api.py | MdMessage._set_flag | def _set_flag(self, flag):
"""Turns the specified flag on"""
self.folder._invalidate_cache()
# TODO::: turn the flag off when it's already on
def replacer(m):
return "%s/%s.%s%s" % (
joinpath(self.folder.base, self.folder.folder, "cur"),
m.grou... | python | def _set_flag(self, flag):
"""Turns the specified flag on"""
self.folder._invalidate_cache()
# TODO::: turn the flag off when it's already on
def replacer(m):
return "%s/%s.%s%s" % (
joinpath(self.folder.base, self.folder.folder, "cur"),
m.grou... | ['def', '_set_flag', '(', 'self', ',', 'flag', ')', ':', 'self', '.', 'folder', '.', '_invalidate_cache', '(', ')', "# TODO::: turn the flag off when it's already on", 'def', 'replacer', '(', 'm', ')', ':', 'return', '"%s/%s.%s%s"', '%', '(', 'joinpath', '(', 'self', '.', 'folder', '.', 'base', ',', 'self', '.', 'folde... | Turns the specified flag on | ['Turns', 'the', 'specified', 'flag', 'on'] | train | https://github.com/nicferrier/md/blob/302ca8882dae060fb15bd5ae470d8e661fb67ec4/src/mdlib/api.py#L143-L159 |
4,507 | ssalentin/plip | plip/modules/supplemental.py | nucleotide_linkage | def nucleotide_linkage(residues):
"""Support for DNA/RNA ligands by finding missing covalent linkages to stitch DNA/RNA together."""
nuc_covalent = []
#######################################
# Basic support for RNA/DNA as ligand #
#######################################
nucleotides = ['A', 'C',... | python | def nucleotide_linkage(residues):
"""Support for DNA/RNA ligands by finding missing covalent linkages to stitch DNA/RNA together."""
nuc_covalent = []
#######################################
# Basic support for RNA/DNA as ligand #
#######################################
nucleotides = ['A', 'C',... | ['def', 'nucleotide_linkage', '(', 'residues', ')', ':', 'nuc_covalent', '=', '[', ']', '#######################################', '# Basic support for RNA/DNA as ligand #', '#######################################', 'nucleotides', '=', '[', "'A'", ',', "'C'", ',', "'T'", ',', "'G'", ',', "'U'", ',', "'DA'", ',', "'DC'... | Support for DNA/RNA ligands by finding missing covalent linkages to stitch DNA/RNA together. | ['Support', 'for', 'DNA', '/', 'RNA', 'ligands', 'by', 'finding', 'missing', 'covalent', 'linkages', 'to', 'stitch', 'DNA', '/', 'RNA', 'together', '.'] | train | https://github.com/ssalentin/plip/blob/906c8d36463689779b403f6c2c9ed06174acaf9a/plip/modules/supplemental.py#L242-L271 |
4,508 | aio-libs/aioredis | aioredis/commands/streams.py | StreamCommandsMixin._xread | def _xread(self, streams, timeout=0, count=None, latest_ids=None):
"""Wraps up common functionality between ``xread()``
and ``xread_group()``
You should probably be using ``xread()`` or ``xread_group()`` directly.
"""
if latest_ids is None:
latest_ids = ['$'] * len(s... | python | def _xread(self, streams, timeout=0, count=None, latest_ids=None):
"""Wraps up common functionality between ``xread()``
and ``xread_group()``
You should probably be using ``xread()`` or ``xread_group()`` directly.
"""
if latest_ids is None:
latest_ids = ['$'] * len(s... | ['def', '_xread', '(', 'self', ',', 'streams', ',', 'timeout', '=', '0', ',', 'count', '=', 'None', ',', 'latest_ids', '=', 'None', ')', ':', 'if', 'latest_ids', 'is', 'None', ':', 'latest_ids', '=', '[', "'$'", ']', '*', 'len', '(', 'streams', ')', 'if', 'len', '(', 'streams', ')', '!=', 'len', '(', 'latest_ids', ')',... | Wraps up common functionality between ``xread()``
and ``xread_group()``
You should probably be using ``xread()`` or ``xread_group()`` directly. | ['Wraps', 'up', 'common', 'functionality', 'between', 'xread', '()', 'and', 'xread_group', '()'] | train | https://github.com/aio-libs/aioredis/blob/e8c33e39558d4cc91cf70dde490d8b330c97dc2e/aioredis/commands/streams.py#L234-L256 |
4,509 | CityOfZion/neo-python-core | neocore/Cryptography/ECCurve.py | EllipticCurve.decompress_from_curve | def decompress_from_curve(self, x, flag):
"""
calculate the y coordinate given only the x value.
there are 2 possible solutions, use 'flag' to select.
"""
cq = self.field.p
x = self.field.value(x)
ysquare = x ** 3 + self.a * x + self.b
ysquare_root = sq... | python | def decompress_from_curve(self, x, flag):
"""
calculate the y coordinate given only the x value.
there are 2 possible solutions, use 'flag' to select.
"""
cq = self.field.p
x = self.field.value(x)
ysquare = x ** 3 + self.a * x + self.b
ysquare_root = sq... | ['def', 'decompress_from_curve', '(', 'self', ',', 'x', ',', 'flag', ')', ':', 'cq', '=', 'self', '.', 'field', '.', 'p', 'x', '=', 'self', '.', 'field', '.', 'value', '(', 'x', ')', 'ysquare', '=', 'x', '**', '3', '+', 'self', '.', 'a', '*', 'x', '+', 'self', '.', 'b', 'ysquare_root', '=', 'sqrtCQ', '(', 'ysquare', '.... | calculate the y coordinate given only the x value.
there are 2 possible solutions, use 'flag' to select. | ['calculate', 'the', 'y', 'coordinate', 'given', 'only', 'the', 'x', 'value', '.', 'there', 'are', '2', 'possible', 'solutions', 'use', 'flag', 'to', 'select', '.'] | train | https://github.com/CityOfZion/neo-python-core/blob/786c02cc2f41712d70b1f064ae3d67f86167107f/neocore/Cryptography/ECCurve.py#L648-L670 |
4,510 | django-treebeard/django-treebeard | treebeard/numconv.py | int2str | def int2str(num, radix=10, alphabet=BASE85):
"""helper function for quick base conversions from integers to strings"""
return NumConv(radix, alphabet).int2str(num) | python | def int2str(num, radix=10, alphabet=BASE85):
"""helper function for quick base conversions from integers to strings"""
return NumConv(radix, alphabet).int2str(num) | ['def', 'int2str', '(', 'num', ',', 'radix', '=', '10', ',', 'alphabet', '=', 'BASE85', ')', ':', 'return', 'NumConv', '(', 'radix', ',', 'alphabet', ')', '.', 'int2str', '(', 'num', ')'] | helper function for quick base conversions from integers to strings | ['helper', 'function', 'for', 'quick', 'base', 'conversions', 'from', 'integers', 'to', 'strings'] | train | https://github.com/django-treebeard/django-treebeard/blob/8042ee939cb45394909237da447f8925e3cc6aa3/treebeard/numconv.py#L108-L110 |
4,511 | sorgerlab/indra | indra/assemblers/pysb/kappa_util.py | im_json_to_graph | def im_json_to_graph(im_json):
"""Return networkx graph from Kappy's influence map JSON.
Parameters
----------
im_json : dict
A JSON dict which contains an influence map generated by Kappy.
Returns
-------
graph : networkx.MultiDiGraph
A graph representing the influence map... | python | def im_json_to_graph(im_json):
"""Return networkx graph from Kappy's influence map JSON.
Parameters
----------
im_json : dict
A JSON dict which contains an influence map generated by Kappy.
Returns
-------
graph : networkx.MultiDiGraph
A graph representing the influence map... | ['def', 'im_json_to_graph', '(', 'im_json', ')', ':', 'imap_data', '=', 'im_json', '[', "'influence map'", ']', '[', "'map'", ']', '# Initialize the graph', 'graph', '=', 'MultiDiGraph', '(', ')', 'id_node_dict', '=', '{', '}', '# Add each node to the graph', 'for', 'node_dict', 'in', 'imap_data', '[', "'nodes'", ']', ... | Return networkx graph from Kappy's influence map JSON.
Parameters
----------
im_json : dict
A JSON dict which contains an influence map generated by Kappy.
Returns
-------
graph : networkx.MultiDiGraph
A graph representing the influence map. | ['Return', 'networkx', 'graph', 'from', 'Kappy', 's', 'influence', 'map', 'JSON', '.'] | train | https://github.com/sorgerlab/indra/blob/79a70415832c5702d7a820c7c9ccc8e25010124b/indra/assemblers/pysb/kappa_util.py#L7-L57 |
4,512 | thekashifmalik/packer | packer/packer.py | Packer.install | def install(cls):
"""Create the required directories in the home directory"""
[os.makedirs('{}/{}'.format(cls.home, cls.dirs[d])) for d in cls.dirs] | python | def install(cls):
"""Create the required directories in the home directory"""
[os.makedirs('{}/{}'.format(cls.home, cls.dirs[d])) for d in cls.dirs] | ['def', 'install', '(', 'cls', ')', ':', '[', 'os', '.', 'makedirs', '(', "'{}/{}'", '.', 'format', '(', 'cls', '.', 'home', ',', 'cls', '.', 'dirs', '[', 'd', ']', ')', ')', 'for', 'd', 'in', 'cls', '.', 'dirs', ']'] | Create the required directories in the home directory | ['Create', 'the', 'required', 'directories', 'in', 'the', 'home', 'directory'] | train | https://github.com/thekashifmalik/packer/blob/736d052d2536ada7733f4b8459e32fb771af2e1c/packer/packer.py#L54-L56 |
4,513 | spyder-ide/spyder | spyder/plugins/help/plugin.py | Help.toggle_wrap_mode | def toggle_wrap_mode(self, checked):
"""Toggle wrap mode"""
self.plain_text.editor.toggle_wrap_mode(checked)
self.set_option('wrap', checked) | python | def toggle_wrap_mode(self, checked):
"""Toggle wrap mode"""
self.plain_text.editor.toggle_wrap_mode(checked)
self.set_option('wrap', checked) | ['def', 'toggle_wrap_mode', '(', 'self', ',', 'checked', ')', ':', 'self', '.', 'plain_text', '.', 'editor', '.', 'toggle_wrap_mode', '(', 'checked', ')', 'self', '.', 'set_option', '(', "'wrap'", ',', 'checked', ')'] | Toggle wrap mode | ['Toggle', 'wrap', 'mode'] | train | https://github.com/spyder-ide/spyder/blob/f76836ce1b924bcc4efd3f74f2960d26a4e528e0/spyder/plugins/help/plugin.py#L341-L344 |
4,514 | sorgerlab/indra | indra/tools/reading/pmid_reading/read_pmids.py | join_json_files | def join_json_files(prefix):
"""Join different REACH output JSON files into a single JSON object.
The output of REACH is broken into three files that need to be joined
before processing. Specifically, there will be three files of the form:
`<prefix>.uaz.<subcategory>.json`.
Parameters
--------... | python | def join_json_files(prefix):
"""Join different REACH output JSON files into a single JSON object.
The output of REACH is broken into three files that need to be joined
before processing. Specifically, there will be three files of the form:
`<prefix>.uaz.<subcategory>.json`.
Parameters
--------... | ['def', 'join_json_files', '(', 'prefix', ')', ':', 'try', ':', 'with', 'open', '(', 'prefix', '+', "'.uaz.entities.json'", ',', "'rt'", ')', 'as', 'f', ':', 'entities', '=', 'json', '.', 'load', '(', 'f', ')', 'with', 'open', '(', 'prefix', '+', "'.uaz.events.json'", ',', "'rt'", ')', 'as', 'f', ':', 'events', '=', 'j... | Join different REACH output JSON files into a single JSON object.
The output of REACH is broken into three files that need to be joined
before processing. Specifically, there will be three files of the form:
`<prefix>.uaz.<subcategory>.json`.
Parameters
----------
prefix : str
The abso... | ['Join', 'different', 'REACH', 'output', 'JSON', 'files', 'into', 'a', 'single', 'JSON', 'object', '.'] | train | https://github.com/sorgerlab/indra/blob/79a70415832c5702d7a820c7c9ccc8e25010124b/indra/tools/reading/pmid_reading/read_pmids.py#L140-L170 |
4,515 | scott-griffiths/bitstring | bitstring.py | Bits._readsie | def _readsie(self, pos):
"""Return interpretation of next bits as a signed interleaved exponential-Golomb code.
Advances position to after the read code.
Raises ReadError if the end of the bitstring is encountered while
reading the code.
"""
codenum, pos = self._readui... | python | def _readsie(self, pos):
"""Return interpretation of next bits as a signed interleaved exponential-Golomb code.
Advances position to after the read code.
Raises ReadError if the end of the bitstring is encountered while
reading the code.
"""
codenum, pos = self._readui... | ['def', '_readsie', '(', 'self', ',', 'pos', ')', ':', 'codenum', ',', 'pos', '=', 'self', '.', '_readuie', '(', 'pos', ')', 'if', 'not', 'codenum', ':', 'return', '0', ',', 'pos', 'try', ':', 'if', 'self', '[', 'pos', ']', ':', 'return', '-', 'codenum', ',', 'pos', '+', '1', 'else', ':', 'return', 'codenum', ',', 'pos... | Return interpretation of next bits as a signed interleaved exponential-Golomb code.
Advances position to after the read code.
Raises ReadError if the end of the bitstring is encountered while
reading the code. | ['Return', 'interpretation', 'of', 'next', 'bits', 'as', 'a', 'signed', 'interleaved', 'exponential', '-', 'Golomb', 'code', '.'] | train | https://github.com/scott-griffiths/bitstring/blob/ab40ae7f0b43fe223a39b63cbc0529b09f3ef653/bitstring.py#L1784-L1802 |
4,516 | eventifyio/eventify | eventify/drivers/crossbar.py | Component.onUserError | def onUserError(self, fail, message):
"""
Handle user errors
"""
self.log.error(fail)
self.log.error(message) | python | def onUserError(self, fail, message):
"""
Handle user errors
"""
self.log.error(fail)
self.log.error(message) | ['def', 'onUserError', '(', 'self', ',', 'fail', ',', 'message', ')', ':', 'self', '.', 'log', '.', 'error', '(', 'fail', ')', 'self', '.', 'log', '.', 'error', '(', 'message', ')'] | Handle user errors | ['Handle', 'user', 'errors'] | train | https://github.com/eventifyio/eventify/blob/0e519964a56bd07a879b266f21f177749c63aaed/eventify/drivers/crossbar.py#L86-L91 |
4,517 | tensorflow/lucid | lucid/misc/gl/meshutil.py | load_obj | def load_obj(fn):
"""Load 3d mesh form .obj' file.
Args:
fn: Input file name or file-like object.
Returns:
dictionary with the following keys (some of which may be missing):
position: np.float32, (n, 3) array, vertex positions
uv: np.float32, (n, 2) array, vertex uv coordinates
n... | python | def load_obj(fn):
"""Load 3d mesh form .obj' file.
Args:
fn: Input file name or file-like object.
Returns:
dictionary with the following keys (some of which may be missing):
position: np.float32, (n, 3) array, vertex positions
uv: np.float32, (n, 2) array, vertex uv coordinates
n... | ['def', 'load_obj', '(', 'fn', ')', ':', 'position', '=', '[', 'np', '.', 'zeros', '(', '3', ',', 'dtype', '=', 'np', '.', 'float32', ')', ']', 'normal', '=', '[', 'np', '.', 'zeros', '(', '3', ',', 'dtype', '=', 'np', '.', 'float32', ')', ']', 'uv', '=', '[', 'np', '.', 'zeros', '(', '2', ',', 'dtype', '=', 'np', '.',... | Load 3d mesh form .obj' file.
Args:
fn: Input file name or file-like object.
Returns:
dictionary with the following keys (some of which may be missing):
position: np.float32, (n, 3) array, vertex positions
uv: np.float32, (n, 2) array, vertex uv coordinates
normal: np.float32, (n, ... | ['Load', '3d', 'mesh', 'form', '.', 'obj', 'file', '.', 'Args', ':', 'fn', ':', 'Input', 'file', 'name', 'or', 'file', '-', 'like', 'object', '.', 'Returns', ':', 'dictionary', 'with', 'the', 'following', 'keys', '(', 'some', 'of', 'which', 'may', 'be', 'missing', ')', ':', 'position', ':', 'np', '.', 'float32', '(', '... | train | https://github.com/tensorflow/lucid/blob/d1a1e2e4fd4be61b89b8cba20dc425a5ae34576e/lucid/misc/gl/meshutil.py#L99-L158 |
4,518 | limodou/uliweb | uliweb/contrib/model_config/__init__.py | get_model | def get_model(sender, model_name, model_inst, model_info, model_config):
"""
#todo Add objcache support
"""
MC = get_mc()
if MC:
model = MC.get((MC.c.model_name==model_name) & (MC.c.uuid!=''))
if model:
cached_inst = __cache__.get(model_name)
if not cached_ins... | python | def get_model(sender, model_name, model_inst, model_info, model_config):
"""
#todo Add objcache support
"""
MC = get_mc()
if MC:
model = MC.get((MC.c.model_name==model_name) & (MC.c.uuid!=''))
if model:
cached_inst = __cache__.get(model_name)
if not cached_ins... | ['def', 'get_model', '(', 'sender', ',', 'model_name', ',', 'model_inst', ',', 'model_info', ',', 'model_config', ')', ':', 'MC', '=', 'get_mc', '(', ')', 'if', 'MC', ':', 'model', '=', 'MC', '.', 'get', '(', '(', 'MC', '.', 'c', '.', 'model_name', '==', 'model_name', ')', '&', '(', 'MC', '.', 'c', '.', 'uuid', '!=', "... | #todo Add objcache support | ['#todo', 'Add', 'objcache', 'support'] | train | https://github.com/limodou/uliweb/blob/34472f25e4bc0b954a35346672f94e84ef18b076/uliweb/contrib/model_config/__init__.py#L37-L67 |
4,519 | prompt-toolkit/pymux | pymux/commands/commands.py | bind_key | def bind_key(pymux, variables):
"""
Bind a key sequence.
-n: Not necessary to use the prefix.
"""
key = variables['<key>']
command = variables['<command>']
arguments = variables['<arguments>']
needs_prefix = not variables['-n']
try:
pymux.key_bindings_manager.add_custom_bind... | python | def bind_key(pymux, variables):
"""
Bind a key sequence.
-n: Not necessary to use the prefix.
"""
key = variables['<key>']
command = variables['<command>']
arguments = variables['<arguments>']
needs_prefix = not variables['-n']
try:
pymux.key_bindings_manager.add_custom_bind... | ['def', 'bind_key', '(', 'pymux', ',', 'variables', ')', ':', 'key', '=', 'variables', '[', "'<key>'", ']', 'command', '=', 'variables', '[', "'<command>'", ']', 'arguments', '=', 'variables', '[', "'<arguments>'", ']', 'needs_prefix', '=', 'not', 'variables', '[', "'-n'", ']', 'try', ':', 'pymux', '.', 'key_bindings_m... | Bind a key sequence.
-n: Not necessary to use the prefix. | ['Bind', 'a', 'key', 'sequence', '.', '-', 'n', ':', 'Not', 'necessary', 'to', 'use', 'the', 'prefix', '.'] | train | https://github.com/prompt-toolkit/pymux/blob/3f66e62b9de4b2251c7f9afad6c516dc5a30ec67/pymux/commands/commands.py#L478-L492 |
4,520 | karel-brinda/rnftools | rnftools/lavender/Bam.py | Bam.create_es | def create_es(self):
"""Create an ES (intermediate) file for this BAM file.
This is the function which asses if an alignment is correct
"""
with (gzip.open(self._es_fn, "tw+") if self.compress_intermediate_files else open(self._es_fn, "w+")) as es_fo:
self.bam2es(
bam_fn... | python | def create_es(self):
"""Create an ES (intermediate) file for this BAM file.
This is the function which asses if an alignment is correct
"""
with (gzip.open(self._es_fn, "tw+") if self.compress_intermediate_files else open(self._es_fn, "w+")) as es_fo:
self.bam2es(
bam_fn... | ['def', 'create_es', '(', 'self', ')', ':', 'with', '(', 'gzip', '.', 'open', '(', 'self', '.', '_es_fn', ',', '"tw+"', ')', 'if', 'self', '.', 'compress_intermediate_files', 'else', 'open', '(', 'self', '.', '_es_fn', ',', '"w+"', ')', ')', 'as', 'es_fo', ':', 'self', '.', 'bam2es', '(', 'bam_fn', '=', 'self', '.', '_... | Create an ES (intermediate) file for this BAM file.
This is the function which asses if an alignment is correct | ['Create', 'an', 'ES', '(', 'intermediate', ')', 'file', 'for', 'this', 'BAM', 'file', '.', 'This', 'is', 'the', 'function', 'which', 'asses', 'if', 'an', 'alignment', 'is', 'correct'] | train | https://github.com/karel-brinda/rnftools/blob/25510798606fbc803a622a1abfcecf06d00d47a9/rnftools/lavender/Bam.py#L239-L249 |
4,521 | johnwmillr/LyricsGenius | lyricsgenius/api.py | API.get_artist | def get_artist(self, id_):
"""Data for a specific artist."""
endpoint = "artists/{id}".format(id=id_)
return self._make_request(endpoint) | python | def get_artist(self, id_):
"""Data for a specific artist."""
endpoint = "artists/{id}".format(id=id_)
return self._make_request(endpoint) | ['def', 'get_artist', '(', 'self', ',', 'id_', ')', ':', 'endpoint', '=', '"artists/{id}"', '.', 'format', '(', 'id', '=', 'id_', ')', 'return', 'self', '.', '_make_request', '(', 'endpoint', ')'] | Data for a specific artist. | ['Data', 'for', 'a', 'specific', 'artist', '.'] | train | https://github.com/johnwmillr/LyricsGenius/blob/e36482f7c42235037f3b9b7013edcd54141124e3/lyricsgenius/api.py#L76-L79 |
4,522 | CivicSpleen/ambry | ambry/bundle/files.py | BuildSourceFileAccessor.objects_to_record | def objects_to_record(self, preference=None):
"""Create file records from objects. """
from ambry.orm.file import File
raise NotImplementedError("Still uses obsolete file_info_map")
for file_const, (file_name, clz) in iteritems(file_info_map):
f = self.file(file_const)
... | python | def objects_to_record(self, preference=None):
"""Create file records from objects. """
from ambry.orm.file import File
raise NotImplementedError("Still uses obsolete file_info_map")
for file_const, (file_name, clz) in iteritems(file_info_map):
f = self.file(file_const)
... | ['def', 'objects_to_record', '(', 'self', ',', 'preference', '=', 'None', ')', ':', 'from', 'ambry', '.', 'orm', '.', 'file', 'import', 'File', 'raise', 'NotImplementedError', '(', '"Still uses obsolete file_info_map"', ')', 'for', 'file_const', ',', '(', 'file_name', ',', 'clz', ')', 'in', 'iteritems', '(', 'file_info... | Create file records from objects. | ['Create', 'file', 'records', 'from', 'objects', '.'] | train | https://github.com/CivicSpleen/ambry/blob/d7f2be4bf1f7ffd086f3fadd4fcae60c32473e42/ambry/bundle/files.py#L1354-L1366 |
4,523 | inasafe/inasafe | safe/gui/tools/options_dialog.py | OptionsDialog.restore_defaults_ratio | def restore_defaults_ratio(self):
"""Restore InaSAFE default ratio."""
# Set the flag to true because user ask to.
self.is_restore_default = True
# remove current default ratio
for i in reversed(list(range(self.container_layout.count()))):
widget = self.container_layo... | python | def restore_defaults_ratio(self):
"""Restore InaSAFE default ratio."""
# Set the flag to true because user ask to.
self.is_restore_default = True
# remove current default ratio
for i in reversed(list(range(self.container_layout.count()))):
widget = self.container_layo... | ['def', 'restore_defaults_ratio', '(', 'self', ')', ':', '# Set the flag to true because user ask to.', 'self', '.', 'is_restore_default', '=', 'True', '# remove current default ratio', 'for', 'i', 'in', 'reversed', '(', 'list', '(', 'range', '(', 'self', '.', 'container_layout', '.', 'count', '(', ')', ')', ')', ')', ... | Restore InaSAFE default ratio. | ['Restore', 'InaSAFE', 'default', 'ratio', '.'] | train | https://github.com/inasafe/inasafe/blob/831d60abba919f6d481dc94a8d988cc205130724/safe/gui/tools/options_dialog.py#L845-L856 |
4,524 | edeposit/edeposit.amqp.ltp | src/edeposit/amqp/ltp/fn_composers.py | _get_suffix | def _get_suffix(path):
"""
Return suffix from `path`.
``/home/xex/somefile.txt`` --> ``txt``.
Args:
path (str): Full file path.
Returns:
str: Suffix.
Raises:
UserWarning: When ``/`` is detected in suffix.
"""
suffix = os.path.basename(path).split(".")[-1]
... | python | def _get_suffix(path):
"""
Return suffix from `path`.
``/home/xex/somefile.txt`` --> ``txt``.
Args:
path (str): Full file path.
Returns:
str: Suffix.
Raises:
UserWarning: When ``/`` is detected in suffix.
"""
suffix = os.path.basename(path).split(".")[-1]
... | ['def', '_get_suffix', '(', 'path', ')', ':', 'suffix', '=', 'os', '.', 'path', '.', 'basename', '(', 'path', ')', '.', 'split', '(', '"."', ')', '[', '-', '1', ']', 'if', '"/"', 'in', 'suffix', ':', 'raise', 'UserWarning', '(', '"Filename can\'t contain \'/\' in suffix (%s)!"', '%', 'path', ')', 'return', 'suffix'] | Return suffix from `path`.
``/home/xex/somefile.txt`` --> ``txt``.
Args:
path (str): Full file path.
Returns:
str: Suffix.
Raises:
UserWarning: When ``/`` is detected in suffix. | ['Return', 'suffix', 'from', 'path', '.'] | train | https://github.com/edeposit/edeposit.amqp.ltp/blob/df9ac7ec6cbdbeaaeed438ca66df75ea967b6d8e/src/edeposit/amqp/ltp/fn_composers.py#L14-L34 |
4,525 | ff0000/scarlet | scarlet/cms/base_views.py | ModelCMSMixin.log_action | def log_action(self, instance, action, action_date=None, url="",
update_parent=True):
"""
Store an action in the database using the CMSLog model.
The following attributes are calculated and set on the log entry:
* **model_repr** - A unicode representation of the inst... | python | def log_action(self, instance, action, action_date=None, url="",
update_parent=True):
"""
Store an action in the database using the CMSLog model.
The following attributes are calculated and set on the log entry:
* **model_repr** - A unicode representation of the inst... | ['def', 'log_action', '(', 'self', ',', 'instance', ',', 'action', ',', 'action_date', '=', 'None', ',', 'url', '=', '""', ',', 'update_parent', '=', 'True', ')', ':', 'section', '=', 'None', 'if', 'self', '.', 'bundle', ':', 'bundle', '=', 'self', '.', 'bundle', 'while', 'bundle', '.', 'parent', ':', 'bundle', '=', 'b... | Store an action in the database using the CMSLog model.
The following attributes are calculated and set on the log entry:
* **model_repr** - A unicode representation of the instance.
* **object_repr** - The verbose_name of the instance model class.
* **section** - The name of ancesto... | ['Store', 'an', 'action', 'in', 'the', 'database', 'using', 'the', 'CMSLog', 'model', '.', 'The', 'following', 'attributes', 'are', 'calculated', 'and', 'set', 'on', 'the', 'log', 'entry', ':'] | train | https://github.com/ff0000/scarlet/blob/6c37befd810916a2d7ffff2cdb2dab57bcb6d12e/scarlet/cms/base_views.py#L469-L528 |
4,526 | chibisov/drf-extensions | docs/backdoc.py | _xml_escape_attr | def _xml_escape_attr(attr, skip_single_quote=True):
"""Escape the given string for use in an HTML/XML tag attribute.
By default this doesn't bother with escaping `'` to `'`, presuming that
the tag attribute is surrounded by double quotes.
"""
escaped = (attr
.replace('&', '&')
... | python | def _xml_escape_attr(attr, skip_single_quote=True):
"""Escape the given string for use in an HTML/XML tag attribute.
By default this doesn't bother with escaping `'` to `'`, presuming that
the tag attribute is surrounded by double quotes.
"""
escaped = (attr
.replace('&', '&')
... | ['def', '_xml_escape_attr', '(', 'attr', ',', 'skip_single_quote', '=', 'True', ')', ':', 'escaped', '=', '(', 'attr', '.', 'replace', '(', "'&'", ',', "'&'", ')', '.', 'replace', '(', '\'"\'', ',', "'"'", ')', '.', 'replace', '(', "'<'", ',', "'<'", ')', '.', 'replace', '(', "'>'", ',', "'>'", ')', ')',... | Escape the given string for use in an HTML/XML tag attribute.
By default this doesn't bother with escaping `'` to `'`, presuming that
the tag attribute is surrounded by double quotes. | ['Escape', 'the', 'given', 'string', 'for', 'use', 'in', 'an', 'HTML', '/', 'XML', 'tag', 'attribute', '.'] | train | https://github.com/chibisov/drf-extensions/blob/1d28a4b28890eab5cd19e93e042f8590c8c2fb8b/docs/backdoc.py#L2164-L2177 |
4,527 | spyder-ide/spyder | spyder/plugins/editor/widgets/editor.py | TabSwitcherWidget.focusOutEvent | def focusOutEvent(self, event):
"""Reimplement Qt method to close the widget when loosing focus."""
event.ignore()
# Inspired from CompletionWidget.focusOutEvent() in file
# widgets/sourcecode/base.py line 212
if sys.platform == "darwin":
if event.reason() != Qt... | python | def focusOutEvent(self, event):
"""Reimplement Qt method to close the widget when loosing focus."""
event.ignore()
# Inspired from CompletionWidget.focusOutEvent() in file
# widgets/sourcecode/base.py line 212
if sys.platform == "darwin":
if event.reason() != Qt... | ['def', 'focusOutEvent', '(', 'self', ',', 'event', ')', ':', 'event', '.', 'ignore', '(', ')', '# Inspired from CompletionWidget.focusOutEvent() in file\r', '# widgets/sourcecode/base.py line 212\r', 'if', 'sys', '.', 'platform', '==', '"darwin"', ':', 'if', 'event', '.', 'reason', '(', ')', '!=', 'Qt', '.', 'ActiveWi... | Reimplement Qt method to close the widget when loosing focus. | ['Reimplement', 'Qt', 'method', 'to', 'close', 'the', 'widget', 'when', 'loosing', 'focus', '.'] | train | https://github.com/spyder-ide/spyder/blob/f76836ce1b924bcc4efd3f74f2960d26a4e528e0/spyder/plugins/editor/widgets/editor.py#L391-L400 |
4,528 | gwpy/gwpy | gwpy/io/cache.py | sieve | def sieve(cache, segment=None):
"""Filter the cache to find those entries that overlap ``segment``
Parameters
----------
cache : `list`
Input list of file paths
segment : `~gwpy.segments.Segment`
The ``[start, stop)`` interval to match against.
"""
return type(cache)(e for ... | python | def sieve(cache, segment=None):
"""Filter the cache to find those entries that overlap ``segment``
Parameters
----------
cache : `list`
Input list of file paths
segment : `~gwpy.segments.Segment`
The ``[start, stop)`` interval to match against.
"""
return type(cache)(e for ... | ['def', 'sieve', '(', 'cache', ',', 'segment', '=', 'None', ')', ':', 'return', 'type', '(', 'cache', ')', '(', 'e', 'for', 'e', 'in', 'cache', 'if', 'segment', '.', 'intersects', '(', 'file_segment', '(', 'e', ')', ')', ')'] | Filter the cache to find those entries that overlap ``segment``
Parameters
----------
cache : `list`
Input list of file paths
segment : `~gwpy.segments.Segment`
The ``[start, stop)`` interval to match against. | ['Filter', 'the', 'cache', 'to', 'find', 'those', 'entries', 'that', 'overlap', 'segment'] | train | https://github.com/gwpy/gwpy/blob/7a92b917e7dd2d99b15895293a1fa1d66cdb210a/gwpy/io/cache.py#L464-L475 |
4,529 | fchorney/rpI2C | rpI2C.py | I2C.write_quick | def write_quick(self):
"""
Send only the read / write bit
"""
self.bus.write_quick(self.address)
self.log.debug("write_quick: Sent the read / write bit") | python | def write_quick(self):
"""
Send only the read / write bit
"""
self.bus.write_quick(self.address)
self.log.debug("write_quick: Sent the read / write bit") | ['def', 'write_quick', '(', 'self', ')', ':', 'self', '.', 'bus', '.', 'write_quick', '(', 'self', '.', 'address', ')', 'self', '.', 'log', '.', 'debug', '(', '"write_quick: Sent the read / write bit"', ')'] | Send only the read / write bit | ['Send', 'only', 'the', 'read', '/', 'write', 'bit'] | train | https://github.com/fchorney/rpI2C/blob/7c60f82fa8c91496a74182373da0b95a13919d6e/rpI2C.py#L39-L44 |
4,530 | saltstack/salt | salt/states/pagerduty_schedule.py | present | def present(profile='pagerduty', subdomain=None, api_key=None, **kwargs):
'''
Ensure that a pagerduty schedule exists.
This method accepts as args everything defined in
https://developer.pagerduty.com/documentation/rest/schedules/create.
This means that most arguments are in a dict called "schedule.... | python | def present(profile='pagerduty', subdomain=None, api_key=None, **kwargs):
'''
Ensure that a pagerduty schedule exists.
This method accepts as args everything defined in
https://developer.pagerduty.com/documentation/rest/schedules/create.
This means that most arguments are in a dict called "schedule.... | ['def', 'present', '(', 'profile', '=', "'pagerduty'", ',', 'subdomain', '=', 'None', ',', 'api_key', '=', 'None', ',', '*', '*', 'kwargs', ')', ':', '# for convenience, we accept id, name, or email as the user id.', 'kwargs', '[', "'schedule'", ']', '[', "'name'", ']', '=', 'kwargs', '[', "'name'", ']', '# match PD AP... | Ensure that a pagerduty schedule exists.
This method accepts as args everything defined in
https://developer.pagerduty.com/documentation/rest/schedules/create.
This means that most arguments are in a dict called "schedule."
User id's can be pagerduty id, or name, or email address. | ['Ensure', 'that', 'a', 'pagerduty', 'schedule', 'exists', '.', 'This', 'method', 'accepts', 'as', 'args', 'everything', 'defined', 'in', 'https', ':', '//', 'developer', '.', 'pagerduty', '.', 'com', '/', 'documentation', '/', 'rest', '/', 'schedules', '/', 'create', '.', 'This', 'means', 'that', 'most', 'arguments', ... | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/states/pagerduty_schedule.py#L48-L77 |
4,531 | jobovy/galpy | galpy/util/bovy_ars.py | bovy_ars | def bovy_ars(domain,isDomainFinite,abcissae,hx,hpx,nsamples=1,
hxparams=(),maxn=100):
"""bovy_ars: Implementation of the Adaptive-Rejection Sampling
algorithm by Gilks & Wild (1992): Adaptive Rejection Sampling
for Gibbs Sampling, Applied Statistics, 41, 337
Based on Wild & Gilks (1993), Al... | python | def bovy_ars(domain,isDomainFinite,abcissae,hx,hpx,nsamples=1,
hxparams=(),maxn=100):
"""bovy_ars: Implementation of the Adaptive-Rejection Sampling
algorithm by Gilks & Wild (1992): Adaptive Rejection Sampling
for Gibbs Sampling, Applied Statistics, 41, 337
Based on Wild & Gilks (1993), Al... | ['def', 'bovy_ars', '(', 'domain', ',', 'isDomainFinite', ',', 'abcissae', ',', 'hx', ',', 'hpx', ',', 'nsamples', '=', '1', ',', 'hxparams', '=', '(', ')', ',', 'maxn', '=', '100', ')', ':', '#First set-up the upper and lower hulls', 'hull', '=', 'setup_hull', '(', 'domain', ',', 'isDomainFinite', ',', 'abcissae', ','... | bovy_ars: Implementation of the Adaptive-Rejection Sampling
algorithm by Gilks & Wild (1992): Adaptive Rejection Sampling
for Gibbs Sampling, Applied Statistics, 41, 337
Based on Wild & Gilks (1993), Algorithm AS 287: Adaptive Rejection
Sampling from Log-concave Density Functions, Applied Statistics, 42... | ['bovy_ars', ':', 'Implementation', 'of', 'the', 'Adaptive', '-', 'Rejection', 'Sampling', 'algorithm', 'by', 'Gilks', '&', 'Wild', '(', '1992', ')', ':', 'Adaptive', 'Rejection', 'Sampling', 'for', 'Gibbs', 'Sampling', 'Applied', 'Statistics', '41', '337', 'Based', 'on', 'Wild', '&', 'Gilks', '(', '1993', ')', 'Algori... | train | https://github.com/jobovy/galpy/blob/9c5b9fe65d58835624dffe432be282060918ee08/galpy/util/bovy_ars.py#L36-L83 |
4,532 | DEIB-GECO/PyGMQL | gmql/dataset/GMQLDataset.py | GMQLDataset.meta_group | def meta_group(self, meta, meta_aggregates=None):
"""
*Wrapper of* ``GROUP``
Group operation only for metadata. For further information check :meth:`~.group`
"""
return self.group(meta=meta, meta_aggregates=meta_aggregates) | python | def meta_group(self, meta, meta_aggregates=None):
"""
*Wrapper of* ``GROUP``
Group operation only for metadata. For further information check :meth:`~.group`
"""
return self.group(meta=meta, meta_aggregates=meta_aggregates) | ['def', 'meta_group', '(', 'self', ',', 'meta', ',', 'meta_aggregates', '=', 'None', ')', ':', 'return', 'self', '.', 'group', '(', 'meta', '=', 'meta', ',', 'meta_aggregates', '=', 'meta_aggregates', ')'] | *Wrapper of* ``GROUP``
Group operation only for metadata. For further information check :meth:`~.group` | ['*', 'Wrapper', 'of', '*', 'GROUP'] | train | https://github.com/DEIB-GECO/PyGMQL/blob/e58b2f9402a86056dcda484a32e3de0bb06ed991/gmql/dataset/GMQLDataset.py#L1345-L1351 |
4,533 | jasonrollins/shareplum | shareplum/shareplum.py | _List.GetAttachmentCollection | def GetAttachmentCollection(self, _id):
"""Get Attachments for given List Item ID"""
# Build Request
soap_request = soap('GetAttachmentCollection')
soap_request.add_parameter('listName', self.listName)
soap_request.add_parameter('listItemID', _id)
self.last_request = str... | python | def GetAttachmentCollection(self, _id):
"""Get Attachments for given List Item ID"""
# Build Request
soap_request = soap('GetAttachmentCollection')
soap_request.add_parameter('listName', self.listName)
soap_request.add_parameter('listItemID', _id)
self.last_request = str... | ['def', 'GetAttachmentCollection', '(', 'self', ',', '_id', ')', ':', '# Build Request', 'soap_request', '=', 'soap', '(', "'GetAttachmentCollection'", ')', 'soap_request', '.', 'add_parameter', '(', "'listName'", ',', 'self', '.', 'listName', ')', 'soap_request', '.', 'add_parameter', '(', "'listItemID'", ',', '_id', ... | Get Attachments for given List Item ID | ['Get', 'Attachments', 'for', 'given', 'List', 'Item', 'ID'] | train | https://github.com/jasonrollins/shareplum/blob/404f320808912619920e2d787f2c4387225a14e0/shareplum/shareplum.py#L706-L731 |
4,534 | dfm/celerite | celerite/celerite.py | GP.log_likelihood | def log_likelihood(self, y, _const=math.log(2.0*math.pi), quiet=False):
"""
Compute the marginalized likelihood of the GP model
The factorized matrix from the previous call to :func:`GP.compute` is
used so ``compute`` must be called first.
Args:
y (array[n]): The ob... | python | def log_likelihood(self, y, _const=math.log(2.0*math.pi), quiet=False):
"""
Compute the marginalized likelihood of the GP model
The factorized matrix from the previous call to :func:`GP.compute` is
used so ``compute`` must be called first.
Args:
y (array[n]): The ob... | ['def', 'log_likelihood', '(', 'self', ',', 'y', ',', '_const', '=', 'math', '.', 'log', '(', '2.0', '*', 'math', '.', 'pi', ')', ',', 'quiet', '=', 'False', ')', ':', 'y', '=', 'self', '.', '_process_input', '(', 'y', ')', 'resid', '=', 'y', '-', 'self', '.', 'mean', '.', 'get_value', '(', 'self', '.', '_t', ')', 'try... | Compute the marginalized likelihood of the GP model
The factorized matrix from the previous call to :func:`GP.compute` is
used so ``compute`` must be called first.
Args:
y (array[n]): The observations at coordinates ``x`` from
:func:`GP.compute`.
quiet (... | ['Compute', 'the', 'marginalized', 'likelihood', 'of', 'the', 'GP', 'model'] | train | https://github.com/dfm/celerite/blob/ad3f471f06b18d233f3dab71bb1c20a316173cae/celerite/celerite.py#L155-L192 |
4,535 | bitcraze/crazyflie-lib-python | lpslib/lopoanchor.py | LoPoAnchor.set_mode | def set_mode(self, anchor_id, mode):
"""
Send a packet to set the anchor mode. If the anchor receive the packet,
it will change mode and resets.
"""
data = struct.pack('<BB', LoPoAnchor.LPP_TYPE_MODE, mode)
self.crazyflie.loc.send_short_lpp_packet(anchor_id, data) | python | def set_mode(self, anchor_id, mode):
"""
Send a packet to set the anchor mode. If the anchor receive the packet,
it will change mode and resets.
"""
data = struct.pack('<BB', LoPoAnchor.LPP_TYPE_MODE, mode)
self.crazyflie.loc.send_short_lpp_packet(anchor_id, data) | ['def', 'set_mode', '(', 'self', ',', 'anchor_id', ',', 'mode', ')', ':', 'data', '=', 'struct', '.', 'pack', '(', "'<BB'", ',', 'LoPoAnchor', '.', 'LPP_TYPE_MODE', ',', 'mode', ')', 'self', '.', 'crazyflie', '.', 'loc', '.', 'send_short_lpp_packet', '(', 'anchor_id', ',', 'data', ')'] | Send a packet to set the anchor mode. If the anchor receive the packet,
it will change mode and resets. | ['Send', 'a', 'packet', 'to', 'set', 'the', 'anchor', 'mode', '.', 'If', 'the', 'anchor', 'receive', 'the', 'packet', 'it', 'will', 'change', 'mode', 'and', 'resets', '.'] | train | https://github.com/bitcraze/crazyflie-lib-python/blob/f6ebb4eb315bbe6e02db518936ac17fb615b2af8/lpslib/lopoanchor.py#L66-L72 |
4,536 | faucamp/python-gsmmodem | gsmmodem/modem.py | GsmModem.sendUssd | def sendUssd(self, ussdString, responseTimeout=15):
""" Starts a USSD session by dialing the the specified USSD string, or \
sends the specified string in the existing USSD session (if any)
:param ussdString: The USSD access number to dial
:param responseTimeout: Maximum... | python | def sendUssd(self, ussdString, responseTimeout=15):
""" Starts a USSD session by dialing the the specified USSD string, or \
sends the specified string in the existing USSD session (if any)
:param ussdString: The USSD access number to dial
:param responseTimeout: Maximum... | ['def', 'sendUssd', '(', 'self', ',', 'ussdString', ',', 'responseTimeout', '=', '15', ')', ':', 'self', '.', '_ussdSessionEvent', '=', 'threading', '.', 'Event', '(', ')', 'try', ':', 'cusdResponse', '=', 'self', '.', 'write', '(', '\'AT+CUSD=1,"{0}",15\'', '.', 'format', '(', 'ussdString', ')', ',', 'timeout', '=', '... | Starts a USSD session by dialing the the specified USSD string, or \
sends the specified string in the existing USSD session (if any)
:param ussdString: The USSD access number to dial
:param responseTimeout: Maximum time to wait a response, in seconds
:raise Tim... | ['Starts', 'a', 'USSD', 'session', 'by', 'dialing', 'the', 'the', 'specified', 'USSD', 'string', 'or', '\\', 'sends', 'the', 'specified', 'string', 'in', 'the', 'existing', 'USSD', 'session', '(', 'if', 'any', ')', ':', 'param', 'ussdString', ':', 'The', 'USSD', 'access', 'number', 'to', 'dial', ':', 'param', 'response... | train | https://github.com/faucamp/python-gsmmodem/blob/834c68b1387ca2c91e2210faa8f75526b39723b5/gsmmodem/modem.py#L668-L699 |
4,537 | saltstack/salt | salt/modules/mac_desktop.py | screensaver | def screensaver():
'''
Launch the screensaver.
CLI Example:
.. code-block:: bash
salt '*' desktop.screensaver
'''
cmd = 'open /System/Library/Frameworks/ScreenSaver.framework/Versions/A/Resources/ScreenSaverEngine.app'
call = __salt__['cmd.run_all'](
cmd,
output_lo... | python | def screensaver():
'''
Launch the screensaver.
CLI Example:
.. code-block:: bash
salt '*' desktop.screensaver
'''
cmd = 'open /System/Library/Frameworks/ScreenSaver.framework/Versions/A/Resources/ScreenSaverEngine.app'
call = __salt__['cmd.run_all'](
cmd,
output_lo... | ['def', 'screensaver', '(', ')', ':', 'cmd', '=', "'open /System/Library/Frameworks/ScreenSaver.framework/Versions/A/Resources/ScreenSaverEngine.app'", 'call', '=', '__salt__', '[', "'cmd.run_all'", ']', '(', 'cmd', ',', 'output_loglevel', '=', "'debug'", ',', 'python_shell', '=', 'False', ')', '_check_cmd', '(', 'call... | Launch the screensaver.
CLI Example:
.. code-block:: bash
salt '*' desktop.screensaver | ['Launch', 'the', 'screensaver', '.'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/mac_desktop.py#L69-L87 |
4,538 | gwastro/pycbc | pycbc/inference/models/marginalized_gaussian_noise.py | MarginalizedGaussianNoise._margtimephase_loglr | def _margtimephase_loglr(self, mf_snr, opt_snr):
"""Returns the log likelihood ratio marginalized over time and phase.
"""
return special.logsumexp(numpy.log(special.i0(mf_snr)),
b=self._deltat) - 0.5*opt_snr | python | def _margtimephase_loglr(self, mf_snr, opt_snr):
"""Returns the log likelihood ratio marginalized over time and phase.
"""
return special.logsumexp(numpy.log(special.i0(mf_snr)),
b=self._deltat) - 0.5*opt_snr | ['def', '_margtimephase_loglr', '(', 'self', ',', 'mf_snr', ',', 'opt_snr', ')', ':', 'return', 'special', '.', 'logsumexp', '(', 'numpy', '.', 'log', '(', 'special', '.', 'i0', '(', 'mf_snr', ')', ')', ',', 'b', '=', 'self', '.', '_deltat', ')', '-', '0.5', '*', 'opt_snr'] | Returns the log likelihood ratio marginalized over time and phase. | ['Returns', 'the', 'log', 'likelihood', 'ratio', 'marginalized', 'over', 'time', 'and', 'phase', '.'] | train | https://github.com/gwastro/pycbc/blob/7a64cdd104d263f1b6ea0b01e6841837d05a4cb3/pycbc/inference/models/marginalized_gaussian_noise.py#L549-L553 |
4,539 | go-macaroon-bakery/py-macaroon-bakery | macaroonbakery/httpbakery/_client.py | _prepare_discharge_hook | def _prepare_discharge_hook(req, client):
''' Return the hook function (called when the response is received.)
This allows us to intercept the response and do any necessary
macaroon discharge before returning.
'''
class Retry:
# Define a local class so that we can use its class variable as
... | python | def _prepare_discharge_hook(req, client):
''' Return the hook function (called when the response is received.)
This allows us to intercept the response and do any necessary
macaroon discharge before returning.
'''
class Retry:
# Define a local class so that we can use its class variable as
... | ['def', '_prepare_discharge_hook', '(', 'req', ',', 'client', ')', ':', 'class', 'Retry', ':', '# Define a local class so that we can use its class variable as', '# mutable state accessed by the closures below.', 'count', '=', '0', 'def', 'hook', '(', 'response', ',', '*', 'args', ',', '*', '*', 'kwargs', ')', ':', "''... | Return the hook function (called when the response is received.)
This allows us to intercept the response and do any necessary
macaroon discharge before returning. | ['Return', 'the', 'hook', 'function', '(', 'called', 'when', 'the', 'response', 'is', 'received', '.', ')'] | train | https://github.com/go-macaroon-bakery/py-macaroon-bakery/blob/63ce1ef1dabe816eb8aaec48fbb46761c34ddf77/macaroonbakery/httpbakery/_client.py#L259-L299 |
4,540 | saltstack/salt | salt/modules/rabbitmq.py | _safe_output | def _safe_output(line):
'''
Looks for rabbitmqctl warning, or general formatting, strings that aren't
intended to be parsed as output.
Returns a boolean whether the line can be parsed as rabbitmqctl output.
'''
return not any([
line.startswith('Listing') and line.endswith('...'),
... | python | def _safe_output(line):
'''
Looks for rabbitmqctl warning, or general formatting, strings that aren't
intended to be parsed as output.
Returns a boolean whether the line can be parsed as rabbitmqctl output.
'''
return not any([
line.startswith('Listing') and line.endswith('...'),
... | ['def', '_safe_output', '(', 'line', ')', ':', 'return', 'not', 'any', '(', '[', 'line', '.', 'startswith', '(', "'Listing'", ')', 'and', 'line', '.', 'endswith', '(', "'...'", ')', ',', 'line', '.', 'startswith', '(', "'Listing'", ')', 'and', "'\\t'", 'not', 'in', 'line', ',', "'...done'", 'in', 'line', ',', 'line', '... | Looks for rabbitmqctl warning, or general formatting, strings that aren't
intended to be parsed as output.
Returns a boolean whether the line can be parsed as rabbitmqctl output. | ['Looks', 'for', 'rabbitmqctl', 'warning', 'or', 'general', 'formatting', 'strings', 'that', 'aren', 't', 'intended', 'to', 'be', 'parsed', 'as', 'output', '.', 'Returns', 'a', 'boolean', 'whether', 'the', 'line', 'can', 'be', 'parsed', 'as', 'rabbitmqctl', 'output', '.'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/rabbitmq.py#L132-L143 |
4,541 | theislab/scanpy | scanpy/tools/_sim.py | GRNsim.set_coupl_old | def set_coupl_old(self):
""" Using the adjacency matrix, sample a coupling matrix.
"""
if self.model == 'krumsiek11' or self.model == 'var':
# we already built the coupling matrix in set_coupl20()
return
self.Coupl = np.zeros((self.dim,self.dim))
for i in ... | python | def set_coupl_old(self):
""" Using the adjacency matrix, sample a coupling matrix.
"""
if self.model == 'krumsiek11' or self.model == 'var':
# we already built the coupling matrix in set_coupl20()
return
self.Coupl = np.zeros((self.dim,self.dim))
for i in ... | ['def', 'set_coupl_old', '(', 'self', ')', ':', 'if', 'self', '.', 'model', '==', "'krumsiek11'", 'or', 'self', '.', 'model', '==', "'var'", ':', '# we already built the coupling matrix in set_coupl20()', 'return', 'self', '.', 'Coupl', '=', 'np', '.', 'zeros', '(', '(', 'self', '.', 'dim', ',', 'self', '.', 'dim', ')'... | Using the adjacency matrix, sample a coupling matrix. | ['Using', 'the', 'adjacency', 'matrix', 'sample', 'a', 'coupling', 'matrix', '.'] | train | https://github.com/theislab/scanpy/blob/9e4e5ee02e04cf618872d9b098e24f0542e8b227/scanpy/tools/_sim.py#L619-L650 |
4,542 | internetarchive/doublethink | doublethink/orm.py | Document.table_ensure | def table_ensure(cls, rr):
'''
Creates the table if it doesn't exist.
'''
dbs = rr.db_list().run()
if not rr.dbname in dbs:
logging.info('creating rethinkdb database %s', repr(rr.dbname))
rr.db_create(rr.dbname).run()
tables = rr.table_list().run()... | python | def table_ensure(cls, rr):
'''
Creates the table if it doesn't exist.
'''
dbs = rr.db_list().run()
if not rr.dbname in dbs:
logging.info('creating rethinkdb database %s', repr(rr.dbname))
rr.db_create(rr.dbname).run()
tables = rr.table_list().run()... | ['def', 'table_ensure', '(', 'cls', ',', 'rr', ')', ':', 'dbs', '=', 'rr', '.', 'db_list', '(', ')', '.', 'run', '(', ')', 'if', 'not', 'rr', '.', 'dbname', 'in', 'dbs', ':', 'logging', '.', 'info', '(', "'creating rethinkdb database %s'", ',', 'repr', '(', 'rr', '.', 'dbname', ')', ')', 'rr', '.', 'db_create', '(', 'r... | Creates the table if it doesn't exist. | ['Creates', 'the', 'table', 'if', 'it', 'doesn', 't', 'exist', '.'] | train | https://github.com/internetarchive/doublethink/blob/f7fc7da725c9b572d473c717b3dad9af98a7a2b4/doublethink/orm.py#L182-L195 |
4,543 | project-ncl/pnc-cli | pnc_cli/projects.py | get_project | def get_project(id=None, name=None):
"""
Get a specific Project by ID or name
"""
content = get_project_raw(id, name)
if content:
return utils.format_json(content) | python | def get_project(id=None, name=None):
"""
Get a specific Project by ID or name
"""
content = get_project_raw(id, name)
if content:
return utils.format_json(content) | ['def', 'get_project', '(', 'id', '=', 'None', ',', 'name', '=', 'None', ')', ':', 'content', '=', 'get_project_raw', '(', 'id', ',', 'name', ')', 'if', 'content', ':', 'return', 'utils', '.', 'format_json', '(', 'content', ')'] | Get a specific Project by ID or name | ['Get', 'a', 'specific', 'Project', 'by', 'ID', 'or', 'name'] | train | https://github.com/project-ncl/pnc-cli/blob/3dc149bf84928f60a8044ac50b58bbaddd451902/pnc_cli/projects.py#L69-L75 |
4,544 | google/grr | grr/server/grr_response_server/databases/mysql_users.py | MySQLDBUsersMixin.ReadGRRUser | def ReadGRRUser(self, username, cursor=None):
"""Reads a user object corresponding to a given name."""
cursor.execute(
"SELECT username, password, ui_mode, canary_mode, user_type "
"FROM grr_users WHERE username_hash = %s", [mysql_utils.Hash(username)])
row = cursor.fetchone()
if row is... | python | def ReadGRRUser(self, username, cursor=None):
"""Reads a user object corresponding to a given name."""
cursor.execute(
"SELECT username, password, ui_mode, canary_mode, user_type "
"FROM grr_users WHERE username_hash = %s", [mysql_utils.Hash(username)])
row = cursor.fetchone()
if row is... | ['def', 'ReadGRRUser', '(', 'self', ',', 'username', ',', 'cursor', '=', 'None', ')', ':', 'cursor', '.', 'execute', '(', '"SELECT username, password, ui_mode, canary_mode, user_type "', '"FROM grr_users WHERE username_hash = %s"', ',', '[', 'mysql_utils', '.', 'Hash', '(', 'username', ')', ']', ')', 'row', '=', 'curso... | Reads a user object corresponding to a given name. | ['Reads', 'a', 'user', 'object', 'corresponding', 'to', 'a', 'given', 'name', '.'] | train | https://github.com/google/grr/blob/5cef4e8e2f0d5df43ea4877e9c798e0bf60bfe74/grr/server/grr_response_server/databases/mysql_users.py#L106-L116 |
4,545 | fastai/fastai | fastai/vision/image.py | Image.refresh | def refresh(self)->None:
"Apply any logit, flow, or affine transfers that have been sent to the `Image`."
if self._logit_px is not None:
self._px = self._logit_px.sigmoid_()
self._logit_px = None
if self._affine_mat is not None or self._flow is not None:
self.... | python | def refresh(self)->None:
"Apply any logit, flow, or affine transfers that have been sent to the `Image`."
if self._logit_px is not None:
self._px = self._logit_px.sigmoid_()
self._logit_px = None
if self._affine_mat is not None or self._flow is not None:
self.... | ['def', 'refresh', '(', 'self', ')', '->', 'None', ':', 'if', 'self', '.', '_logit_px', 'is', 'not', 'None', ':', 'self', '.', '_px', '=', 'self', '.', '_logit_px', '.', 'sigmoid_', '(', ')', 'self', '.', '_logit_px', '=', 'None', 'if', 'self', '.', '_affine_mat', 'is', 'not', 'None', 'or', 'self', '.', '_flow', 'is', ... | Apply any logit, flow, or affine transfers that have been sent to the `Image`. | ['Apply', 'any', 'logit', 'flow', 'or', 'affine', 'transfers', 'that', 'have', 'been', 'sent', 'to', 'the', 'Image', '.'] | train | https://github.com/fastai/fastai/blob/9fb84a5cdefe5a766cdb792b8f5d8971737b7e67/fastai/vision/image.py#L126-L135 |
4,546 | Alignak-monitoring/alignak | alignak/external_command.py | ExternalCommandManager.enable_host_svc_notifications | def enable_host_svc_notifications(self, host):
"""Enable services notifications for a host
Format of the line that triggers function call::
ENABLE_HOST_SVC_NOTIFICATIONS;<host_name>
:param host: host to edit
:type host: alignak.objects.host.Host
:return: None
""... | python | def enable_host_svc_notifications(self, host):
"""Enable services notifications for a host
Format of the line that triggers function call::
ENABLE_HOST_SVC_NOTIFICATIONS;<host_name>
:param host: host to edit
:type host: alignak.objects.host.Host
:return: None
""... | ['def', 'enable_host_svc_notifications', '(', 'self', ',', 'host', ')', ':', 'for', 'service_id', 'in', 'host', '.', 'services', ':', 'if', 'service_id', 'in', 'self', '.', 'daemon', '.', 'services', ':', 'service', '=', 'self', '.', 'daemon', '.', 'services', '[', 'service_id', ']', 'self', '.', 'enable_svc_notificati... | Enable services notifications for a host
Format of the line that triggers function call::
ENABLE_HOST_SVC_NOTIFICATIONS;<host_name>
:param host: host to edit
:type host: alignak.objects.host.Host
:return: None | ['Enable', 'services', 'notifications', 'for', 'a', 'host', 'Format', 'of', 'the', 'line', 'that', 'triggers', 'function', 'call', '::'] | train | https://github.com/Alignak-monitoring/alignak/blob/f3c145207e83159b799d3714e4241399c7740a64/alignak/external_command.py#L2821-L2835 |
4,547 | thiagopbueno/rddl2tf | rddl2tf/compiler.py | Compiler.compile_action_bound_constraints | def compile_action_bound_constraints(self,
state: Sequence[tf.Tensor]) -> Dict[str, Bounds]:
'''Compiles all actions bounds for the given `state`.
Args:
state (Sequence[tf.Tensor]): The current state fluents.
Returns:
A mapping from action names to a pair of... | python | def compile_action_bound_constraints(self,
state: Sequence[tf.Tensor]) -> Dict[str, Bounds]:
'''Compiles all actions bounds for the given `state`.
Args:
state (Sequence[tf.Tensor]): The current state fluents.
Returns:
A mapping from action names to a pair of... | ['def', 'compile_action_bound_constraints', '(', 'self', ',', 'state', ':', 'Sequence', '[', 'tf', '.', 'Tensor', ']', ')', '->', 'Dict', '[', 'str', ',', 'Bounds', ']', ':', 'scope', '=', 'self', '.', 'action_precondition_scope', '(', 'state', ')', 'lower_bounds', '=', 'self', '.', 'rddl', '.', 'domain', '.', 'action_... | Compiles all actions bounds for the given `state`.
Args:
state (Sequence[tf.Tensor]): The current state fluents.
Returns:
A mapping from action names to a pair of
:obj:`rddl2tf.fluent.TensorFluent` representing
its lower and upper bounds. | ['Compiles', 'all', 'actions', 'bounds', 'for', 'the', 'given', 'state', '.'] | train | https://github.com/thiagopbueno/rddl2tf/blob/f7c03d3a74d2663807c1e23e04eeed2e85166b71/rddl2tf/compiler.py#L340-L377 |
4,548 | MonashBI/arcana | arcana/utils/base.py | iscontainer | def iscontainer(*items):
"""
Checks whether all the provided items are containers (i.e of class list,
dict, tuple, etc...)
"""
return all(isinstance(i, Iterable) and not isinstance(i, basestring)
for i in items) | python | def iscontainer(*items):
"""
Checks whether all the provided items are containers (i.e of class list,
dict, tuple, etc...)
"""
return all(isinstance(i, Iterable) and not isinstance(i, basestring)
for i in items) | ['def', 'iscontainer', '(', '*', 'items', ')', ':', 'return', 'all', '(', 'isinstance', '(', 'i', ',', 'Iterable', ')', 'and', 'not', 'isinstance', '(', 'i', ',', 'basestring', ')', 'for', 'i', 'in', 'items', ')'] | Checks whether all the provided items are containers (i.e of class list,
dict, tuple, etc...) | ['Checks', 'whether', 'all', 'the', 'provided', 'items', 'are', 'containers', '(', 'i', '.', 'e', 'of', 'class', 'list', 'dict', 'tuple', 'etc', '...', ')'] | train | https://github.com/MonashBI/arcana/blob/d6271a29d13733d00422d11417af8d200be62acc/arcana/utils/base.py#L161-L167 |
4,549 | materialsproject/pymatgen | pymatgen/analysis/chemenv/coordination_environments/structure_environments.py | ChemicalEnvironments.minimum_geometries | def minimum_geometries(self, n=None, symmetry_measure_type=None, max_csm=None):
"""
Returns a list of geometries with increasing continuous symmetry measure in this ChemicalEnvironments object
:param n: Number of geometries to be included in the list
:return: list of geometries with incr... | python | def minimum_geometries(self, n=None, symmetry_measure_type=None, max_csm=None):
"""
Returns a list of geometries with increasing continuous symmetry measure in this ChemicalEnvironments object
:param n: Number of geometries to be included in the list
:return: list of geometries with incr... | ['def', 'minimum_geometries', '(', 'self', ',', 'n', '=', 'None', ',', 'symmetry_measure_type', '=', 'None', ',', 'max_csm', '=', 'None', ')', ':', 'cglist', '=', '[', 'cg', 'for', 'cg', 'in', 'self', '.', 'coord_geoms', ']', 'if', 'symmetry_measure_type', 'is', 'None', ':', 'csms', '=', 'np', '.', 'array', '(', '[', '... | Returns a list of geometries with increasing continuous symmetry measure in this ChemicalEnvironments object
:param n: Number of geometries to be included in the list
:return: list of geometries with increasing continuous symmetry measure in this ChemicalEnvironments object
:raise: ValueError if... | ['Returns', 'a', 'list', 'of', 'geometries', 'with', 'increasing', 'continuous', 'symmetry', 'measure', 'in', 'this', 'ChemicalEnvironments', 'object', ':', 'param', 'n', ':', 'Number', 'of', 'geometries', 'to', 'be', 'included', 'in', 'the', 'list', ':', 'return', ':', 'list', 'of', 'geometries', 'with', 'increasing',... | train | https://github.com/materialsproject/pymatgen/blob/4ca558cf72f8d5f8a1f21dfdfc0181a971c186da/pymatgen/analysis/chemenv/coordination_environments/structure_environments.py#L1571-L1594 |
4,550 | horejsek/python-webdriverwrapper | webdriverwrapper/wrapper.py | _WebdriverWrapper.close_window | def close_window(self, window_name=None, title=None, url=None):
"""
WebDriver implements only closing current window. If you want to close
some window without having to switch to it, use this method.
"""
main_window_handle = self.current_window_handle
self.switch_to_windo... | python | def close_window(self, window_name=None, title=None, url=None):
"""
WebDriver implements only closing current window. If you want to close
some window without having to switch to it, use this method.
"""
main_window_handle = self.current_window_handle
self.switch_to_windo... | ['def', 'close_window', '(', 'self', ',', 'window_name', '=', 'None', ',', 'title', '=', 'None', ',', 'url', '=', 'None', ')', ':', 'main_window_handle', '=', 'self', '.', 'current_window_handle', 'self', '.', 'switch_to_window', '(', 'window_name', ',', 'title', ',', 'url', ')', 'self', '.', 'close', '(', ')', 'self',... | WebDriver implements only closing current window. If you want to close
some window without having to switch to it, use this method. | ['WebDriver', 'implements', 'only', 'closing', 'current', 'window', '.', 'If', 'you', 'want', 'to', 'close', 'some', 'window', 'without', 'having', 'to', 'switch', 'to', 'it', 'use', 'this', 'method', '.'] | train | https://github.com/horejsek/python-webdriverwrapper/blob/a492f79ab60ed83d860dd817b6a0961500d7e3f5/webdriverwrapper/wrapper.py#L516-L524 |
4,551 | teepark/greenhouse | greenhouse/backdoor.py | backdoor_handler | def backdoor_handler(clientsock, namespace=None):
"""start an interactive python interpreter on an existing connection
.. note::
this function will block for as long as the connection remains alive.
:param sock: the socket on which to serve the interpreter
:type sock: :class:`Socket<greenhouse... | python | def backdoor_handler(clientsock, namespace=None):
"""start an interactive python interpreter on an existing connection
.. note::
this function will block for as long as the connection remains alive.
:param sock: the socket on which to serve the interpreter
:type sock: :class:`Socket<greenhouse... | ['def', 'backdoor_handler', '(', 'clientsock', ',', 'namespace', '=', 'None', ')', ':', 'namespace', '=', '{', '}', 'if', 'namespace', 'is', 'None', 'else', 'namespace', '.', 'copy', '(', ')', 'console', '=', 'code', '.', 'InteractiveConsole', '(', 'namespace', ')', 'multiline_statement', '=', '[', ']', 'stdout', ',', ... | start an interactive python interpreter on an existing connection
.. note::
this function will block for as long as the connection remains alive.
:param sock: the socket on which to serve the interpreter
:type sock: :class:`Socket<greenhouse.io.sockets.Socket>`
:param namespace:
the lo... | ['start', 'an', 'interactive', 'python', 'interpreter', 'on', 'an', 'existing', 'connection'] | train | https://github.com/teepark/greenhouse/blob/8fd1be4f5443ba090346b5ec82fdbeb0a060d956/greenhouse/backdoor.py#L70-L112 |
4,552 | linuxsoftware/ls.joyous | ls/joyous/models/events.py | EventBase._removeContentPanels | def _removeContentPanels(cls, remove):
"""
Remove the panels and so hide the fields named.
"""
if type(remove) is str:
remove = [remove]
cls.content_panels = [panel for panel in cls.content_panels
if getattr(panel, "field_name", None) not... | python | def _removeContentPanels(cls, remove):
"""
Remove the panels and so hide the fields named.
"""
if type(remove) is str:
remove = [remove]
cls.content_panels = [panel for panel in cls.content_panels
if getattr(panel, "field_name", None) not... | ['def', '_removeContentPanels', '(', 'cls', ',', 'remove', ')', ':', 'if', 'type', '(', 'remove', ')', 'is', 'str', ':', 'remove', '=', '[', 'remove', ']', 'cls', '.', 'content_panels', '=', '[', 'panel', 'for', 'panel', 'in', 'cls', '.', 'content_panels', 'if', 'getattr', '(', 'panel', ',', '"field_name"', ',', 'None'... | Remove the panels and so hide the fields named. | ['Remove', 'the', 'panels', 'and', 'so', 'hide', 'the', 'fields', 'named', '.'] | train | https://github.com/linuxsoftware/ls.joyous/blob/316283140ca5171a68ad3170a5964fdc89be0b56/ls/joyous/models/events.py#L553-L560 |
4,553 | saltstack/salt | salt/modules/boto_ec2.py | disassociate_eip_address | def disassociate_eip_address(public_ip=None, association_id=None, region=None,
key=None, keyid=None, profile=None):
'''
Disassociate an Elastic IP address from a currently running instance. This
requires exactly one of either 'association_id' or 'public_ip', depending
on whe... | python | def disassociate_eip_address(public_ip=None, association_id=None, region=None,
key=None, keyid=None, profile=None):
'''
Disassociate an Elastic IP address from a currently running instance. This
requires exactly one of either 'association_id' or 'public_ip', depending
on whe... | ['def', 'disassociate_eip_address', '(', 'public_ip', '=', 'None', ',', 'association_id', '=', 'None', ',', 'region', '=', 'None', ',', 'key', '=', 'None', ',', 'keyid', '=', 'None', ',', 'profile', '=', 'None', ')', ':', 'conn', '=', '_get_conn', '(', 'region', '=', 'region', ',', 'key', '=', 'key', ',', 'keyid', '=',... | Disassociate an Elastic IP address from a currently running instance. This
requires exactly one of either 'association_id' or 'public_ip', depending
on whether you’re dealing with a VPC or EC2 Classic address.
public_ip
(string) – Public IP address, for EC2 Classic allocations.
association_id
... | ['Disassociate', 'an', 'Elastic', 'IP', 'address', 'from', 'a', 'currently', 'running', 'instance', '.', 'This', 'requires', 'exactly', 'one', 'of', 'either', 'association_id', 'or', 'public_ip', 'depending', 'on', 'whether', 'you’re', 'dealing', 'with', 'a', 'VPC', 'or', 'EC2', 'Classic', 'address', '.'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/modules/boto_ec2.py#L396-L425 |
4,554 | sosreport/sos | sos/plugins/__init__.py | SCLPlugin.add_copy_spec_scl | def add_copy_spec_scl(self, scl, copyspecs):
"""Same as add_copy_spec, except that it prepends path to SCL root
to "copyspecs".
"""
if isinstance(copyspecs, six.string_types):
copyspecs = [copyspecs]
scl_copyspecs = []
for copyspec in copyspecs:
sc... | python | def add_copy_spec_scl(self, scl, copyspecs):
"""Same as add_copy_spec, except that it prepends path to SCL root
to "copyspecs".
"""
if isinstance(copyspecs, six.string_types):
copyspecs = [copyspecs]
scl_copyspecs = []
for copyspec in copyspecs:
sc... | ['def', 'add_copy_spec_scl', '(', 'self', ',', 'scl', ',', 'copyspecs', ')', ':', 'if', 'isinstance', '(', 'copyspecs', ',', 'six', '.', 'string_types', ')', ':', 'copyspecs', '=', '[', 'copyspecs', ']', 'scl_copyspecs', '=', '[', ']', 'for', 'copyspec', 'in', 'copyspecs', ':', 'scl_copyspecs', '.', 'append', '(', 'sel... | Same as add_copy_spec, except that it prepends path to SCL root
to "copyspecs". | ['Same', 'as', 'add_copy_spec', 'except', 'that', 'it', 'prepends', 'path', 'to', 'SCL', 'root', 'to', 'copyspecs', '.'] | train | https://github.com/sosreport/sos/blob/2ebc04da53dc871c8dd5243567afa4f8592dca29/sos/plugins/__init__.py#L1492-L1501 |
4,555 | CamDavidsonPilon/lifelines | lifelines/fitters/__init__.py | ParametericUnivariateFitter.confidence_interval_hazard_ | def confidence_interval_hazard_(self):
"""
The confidence interval of the hazard.
"""
return self._compute_confidence_bounds_of_transform(self._hazard, self.alpha, self._ci_labels) | python | def confidence_interval_hazard_(self):
"""
The confidence interval of the hazard.
"""
return self._compute_confidence_bounds_of_transform(self._hazard, self.alpha, self._ci_labels) | ['def', 'confidence_interval_hazard_', '(', 'self', ')', ':', 'return', 'self', '.', '_compute_confidence_bounds_of_transform', '(', 'self', '.', '_hazard', ',', 'self', '.', 'alpha', ',', 'self', '.', '_ci_labels', ')'] | The confidence interval of the hazard. | ['The', 'confidence', 'interval', 'of', 'the', 'hazard', '.'] | train | https://github.com/CamDavidsonPilon/lifelines/blob/bdf6be6f1d10eea4c46365ee0ee6a47d8c30edf8/lifelines/fitters/__init__.py#L1032-L1036 |
4,556 | vinci1it2000/schedula | schedula/utils/alg.py | remove_edge_fun | def remove_edge_fun(graph):
"""
Returns a function that removes an edge from the `graph`.
..note:: The out node is removed if this is isolate.
:param graph:
A directed graph.
:type graph: networkx.classes.digraph.DiGraph
:return:
A function that remove an edge from the `graph`... | python | def remove_edge_fun(graph):
"""
Returns a function that removes an edge from the `graph`.
..note:: The out node is removed if this is isolate.
:param graph:
A directed graph.
:type graph: networkx.classes.digraph.DiGraph
:return:
A function that remove an edge from the `graph`... | ['def', 'remove_edge_fun', '(', 'graph', ')', ':', '# Namespace shortcut for speed.', 'rm_edge', ',', 'rm_node', '=', 'graph', '.', 'remove_edge', ',', 'graph', '.', 'remove_node', 'from', 'networkx', 'import', 'is_isolate', 'def', 'remove_edge', '(', 'u', ',', 'v', ')', ':', 'rm_edge', '(', 'u', ',', 'v', ')', '# Remo... | Returns a function that removes an edge from the `graph`.
..note:: The out node is removed if this is isolate.
:param graph:
A directed graph.
:type graph: networkx.classes.digraph.DiGraph
:return:
A function that remove an edge from the `graph`.
:rtype: callable | ['Returns', 'a', 'function', 'that', 'removes', 'an', 'edge', 'from', 'the', 'graph', '.'] | train | https://github.com/vinci1it2000/schedula/blob/addb9fd685be81544b796c51383ac00a31543ce9/schedula/utils/alg.py#L48-L72 |
4,557 | mlperf/training | rnn_translator/pytorch/seq2seq/models/decoder.py | ResidualRecurrentDecoder.package_hidden | def package_hidden(self):
"""
Flattens the hidden state from all LSTM layers into one tensor (for
the sequence generator).
"""
if self.inference:
hidden = torch.cat(tuple(itertools.chain(*self.next_hidden)))
else:
hidden = None
return hidde... | python | def package_hidden(self):
"""
Flattens the hidden state from all LSTM layers into one tensor (for
the sequence generator).
"""
if self.inference:
hidden = torch.cat(tuple(itertools.chain(*self.next_hidden)))
else:
hidden = None
return hidde... | ['def', 'package_hidden', '(', 'self', ')', ':', 'if', 'self', '.', 'inference', ':', 'hidden', '=', 'torch', '.', 'cat', '(', 'tuple', '(', 'itertools', '.', 'chain', '(', '*', 'self', '.', 'next_hidden', ')', ')', ')', 'else', ':', 'hidden', '=', 'None', 'return', 'hidden'] | Flattens the hidden state from all LSTM layers into one tensor (for
the sequence generator). | ['Flattens', 'the', 'hidden', 'state', 'from', 'all', 'LSTM', 'layers', 'into', 'one', 'tensor', '(', 'for', 'the', 'sequence', 'generator', ')', '.'] | train | https://github.com/mlperf/training/blob/1c6ae725a81d15437a2b2df05cac0673fde5c3a4/rnn_translator/pytorch/seq2seq/models/decoder.py#L176-L185 |
4,558 | RomelTorres/alpha_vantage | alpha_vantage/alphavantage.py | AlphaVantage.map_to_matype | def map_to_matype(self, matype):
""" Convert to the alpha vantage math type integer. It returns an
integer correspondent to the type of math to apply to a function. It
raises ValueError if an integer greater than the supported math types
is given.
Keyword Arguments:
... | python | def map_to_matype(self, matype):
""" Convert to the alpha vantage math type integer. It returns an
integer correspondent to the type of math to apply to a function. It
raises ValueError if an integer greater than the supported math types
is given.
Keyword Arguments:
... | ['def', 'map_to_matype', '(', 'self', ',', 'matype', ')', ':', '# Check if it is an integer or a string', 'try', ':', 'value', '=', 'int', '(', 'matype', ')', 'if', 'abs', '(', 'value', ')', '>', 'len', '(', 'AlphaVantage', '.', '_ALPHA_VANTAGE_MATH_MAP', ')', ':', 'raise', 'ValueError', '(', '"The value {} is not supp... | Convert to the alpha vantage math type integer. It returns an
integer correspondent to the type of math to apply to a function. It
raises ValueError if an integer greater than the supported math types
is given.
Keyword Arguments:
matype: The math type of the alpha vantage a... | ['Convert', 'to', 'the', 'alpha', 'vantage', 'math', 'type', 'integer', '.', 'It', 'returns', 'an', 'integer', 'correspondent', 'to', 'the', 'type', 'of', 'math', 'to', 'apply', 'to', 'a', 'function', '.', 'It', 'raises', 'ValueError', 'if', 'an', 'integer', 'greater', 'than', 'the', 'supported', 'math', 'types', 'is',... | train | https://github.com/RomelTorres/alpha_vantage/blob/4e0b5057e520e3e3de69cf947301765817290121/alpha_vantage/alphavantage.py#L214-L241 |
4,559 | zarr-developers/zarr | zarr/creation.py | open_like | def open_like(a, path, **kwargs):
"""Open a persistent array like `a`."""
_like_args(a, kwargs)
if isinstance(a, Array):
kwargs.setdefault('fill_value', a.fill_value)
return open_array(path, **kwargs) | python | def open_like(a, path, **kwargs):
"""Open a persistent array like `a`."""
_like_args(a, kwargs)
if isinstance(a, Array):
kwargs.setdefault('fill_value', a.fill_value)
return open_array(path, **kwargs) | ['def', 'open_like', '(', 'a', ',', 'path', ',', '*', '*', 'kwargs', ')', ':', '_like_args', '(', 'a', ',', 'kwargs', ')', 'if', 'isinstance', '(', 'a', ',', 'Array', ')', ':', 'kwargs', '.', 'setdefault', '(', "'fill_value'", ',', 'a', '.', 'fill_value', ')', 'return', 'open_array', '(', 'path', ',', '*', '*', 'kwargs... | Open a persistent array like `a`. | ['Open', 'a', 'persistent', 'array', 'like', 'a', '.'] | train | https://github.com/zarr-developers/zarr/blob/fb8e6d5ea6bc26e451e5cf0eaaee36977556d5b5/zarr/creation.py#L538-L543 |
4,560 | Godley/MuseParse | MuseParse/classes/ObjectHierarchy/TreeClasses/PartNode.py | PartNode.toLily | def toLily(self):
'''
Method which converts the object instance, its attributes and children to a string of lilypond code
:return: str of lilypond code
'''
self.CheckDivisions()
self.CheckTotals()
staves = self.GetChildrenIndexes()
name = ""
short... | python | def toLily(self):
'''
Method which converts the object instance, its attributes and children to a string of lilypond code
:return: str of lilypond code
'''
self.CheckDivisions()
self.CheckTotals()
staves = self.GetChildrenIndexes()
name = ""
short... | ['def', 'toLily', '(', 'self', ')', ':', 'self', '.', 'CheckDivisions', '(', ')', 'self', '.', 'CheckTotals', '(', ')', 'staves', '=', 'self', '.', 'GetChildrenIndexes', '(', ')', 'name', '=', '""', 'shortname', '=', '""', 'if', 'hasattr', '(', 'self', '.', 'item', ',', '"name"', ')', ':', 'name', '=', 'self', '.', 'it... | Method which converts the object instance, its attributes and children to a string of lilypond code
:return: str of lilypond code | ['Method', 'which', 'converts', 'the', 'object', 'instance', 'its', 'attributes', 'and', 'children', 'to', 'a', 'string', 'of', 'lilypond', 'code'] | train | https://github.com/Godley/MuseParse/blob/23cecafa1fdc0f2d6a87760553572b459f3c9904/MuseParse/classes/ObjectHierarchy/TreeClasses/PartNode.py#L200-L250 |
4,561 | nicolargo/glances | glances/stats.py | GlancesStats._load_plugin | def _load_plugin(self, plugin_script, args=None, config=None):
"""Load the plugin (script), init it and add to the _plugin dict."""
# The key is the plugin name
# for example, the file glances_xxx.py
# generate self._plugins_list["xxx"] = ...
name = plugin_script[len(self.header)... | python | def _load_plugin(self, plugin_script, args=None, config=None):
"""Load the plugin (script), init it and add to the _plugin dict."""
# The key is the plugin name
# for example, the file glances_xxx.py
# generate self._plugins_list["xxx"] = ...
name = plugin_script[len(self.header)... | ['def', '_load_plugin', '(', 'self', ',', 'plugin_script', ',', 'args', '=', 'None', ',', 'config', '=', 'None', ')', ':', '# The key is the plugin name', '# for example, the file glances_xxx.py', '# generate self._plugins_list["xxx"] = ...', 'name', '=', 'plugin_script', '[', 'len', '(', 'self', '.', 'header', ')', ':... | Load the plugin (script), init it and add to the _plugin dict. | ['Load', 'the', 'plugin', '(', 'script', ')', 'init', 'it', 'and', 'add', 'to', 'the', '_plugin', 'dict', '.'] | train | https://github.com/nicolargo/glances/blob/5bd4d587a736e0d2b03170b56926841d2a3eb7ee/glances/stats.py#L106-L129 |
4,562 | amzn/ion-python | amazon/ion/reader_managed.py | _ManagedContext.resolve | def resolve(self, token):
"""Attempts to resolve the :class:`SymbolToken` against the current table.
If the ``text`` is not None, the token is returned, otherwise, a token
in the table is attempted to be retrieved. If not token is found, then
this method will raise.
"""
... | python | def resolve(self, token):
"""Attempts to resolve the :class:`SymbolToken` against the current table.
If the ``text`` is not None, the token is returned, otherwise, a token
in the table is attempted to be retrieved. If not token is found, then
this method will raise.
"""
... | ['def', 'resolve', '(', 'self', ',', 'token', ')', ':', 'if', 'token', '.', 'text', 'is', 'not', 'None', ':', 'return', 'token', 'resolved_token', '=', 'self', '.', 'symbol_table', '.', 'get', '(', 'token', '.', 'sid', ',', 'None', ')', 'if', 'resolved_token', 'is', 'None', ':', 'raise', 'IonException', '(', "'Out of r... | Attempts to resolve the :class:`SymbolToken` against the current table.
If the ``text`` is not None, the token is returned, otherwise, a token
in the table is attempted to be retrieved. If not token is found, then
this method will raise. | ['Attempts', 'to', 'resolve', 'the', ':', 'class', ':', 'SymbolToken', 'against', 'the', 'current', 'table', '.'] | train | https://github.com/amzn/ion-python/blob/0b21fa3ba7755f55f745e4aa970d86343b82449d/amazon/ion/reader_managed.py#L40-L52 |
4,563 | biolink/biolink-model | metamodel/utils/loadschema.py | DupCheckYamlLoader.map_constructor | def map_constructor(self, loader, node, deep=False):
""" Walk the mapping, recording any duplicate keys.
"""
mapping = {}
for key_node, value_node in node.value:
key = loader.construct_object(key_node, deep=deep)
value = loader.construct_object(value_node, deep=... | python | def map_constructor(self, loader, node, deep=False):
""" Walk the mapping, recording any duplicate keys.
"""
mapping = {}
for key_node, value_node in node.value:
key = loader.construct_object(key_node, deep=deep)
value = loader.construct_object(value_node, deep=... | ['def', 'map_constructor', '(', 'self', ',', 'loader', ',', 'node', ',', 'deep', '=', 'False', ')', ':', 'mapping', '=', '{', '}', 'for', 'key_node', ',', 'value_node', 'in', 'node', '.', 'value', ':', 'key', '=', 'loader', '.', 'construct_object', '(', 'key_node', ',', 'deep', '=', 'deep', ')', 'value', '=', 'loader',... | Walk the mapping, recording any duplicate keys. | ['Walk', 'the', 'mapping', 'recording', 'any', 'duplicate', 'keys', '.'] | train | https://github.com/biolink/biolink-model/blob/f379e28d5d4085e1115798c6cb28e5acc4dba8b4/metamodel/utils/loadschema.py#L69-L81 |
4,564 | pybel/pybel | src/pybel/canonicalize.py | group_citation_edges | def group_citation_edges(edges: Iterable[EdgeTuple]) -> Iterable[Tuple[str, Iterable[EdgeTuple]]]:
"""Return an iterator over pairs of citation values and their corresponding edge iterators."""
return itt.groupby(edges, key=_citation_sort_key) | python | def group_citation_edges(edges: Iterable[EdgeTuple]) -> Iterable[Tuple[str, Iterable[EdgeTuple]]]:
"""Return an iterator over pairs of citation values and their corresponding edge iterators."""
return itt.groupby(edges, key=_citation_sort_key) | ['def', 'group_citation_edges', '(', 'edges', ':', 'Iterable', '[', 'EdgeTuple', ']', ')', '->', 'Iterable', '[', 'Tuple', '[', 'str', ',', 'Iterable', '[', 'EdgeTuple', ']', ']', ']', ':', 'return', 'itt', '.', 'groupby', '(', 'edges', ',', 'key', '=', '_citation_sort_key', ')'] | Return an iterator over pairs of citation values and their corresponding edge iterators. | ['Return', 'an', 'iterator', 'over', 'pairs', 'of', 'citation', 'values', 'and', 'their', 'corresponding', 'edge', 'iterators', '.'] | train | https://github.com/pybel/pybel/blob/c8a7a1bdae4c475fa2a8c77f3a9a5f6d79556ca0/src/pybel/canonicalize.py#L204-L206 |
4,565 | datamachine/twx.botapi | twx/botapi/botapi.py | TelegramBot.set_webhook | def set_webhook(self, *args, **kwargs):
"""See :func:`set_webhook`"""
return set_webhook(*args, **self._merge_overrides(**kwargs)).run() | python | def set_webhook(self, *args, **kwargs):
"""See :func:`set_webhook`"""
return set_webhook(*args, **self._merge_overrides(**kwargs)).run() | ['def', 'set_webhook', '(', 'self', ',', '*', 'args', ',', '*', '*', 'kwargs', ')', ':', 'return', 'set_webhook', '(', '*', 'args', ',', '*', '*', 'self', '.', '_merge_overrides', '(', '*', '*', 'kwargs', ')', ')', '.', 'run', '(', ')'] | See :func:`set_webhook` | ['See', ':', 'func', ':', 'set_webhook'] | train | https://github.com/datamachine/twx.botapi/blob/c85184da738169e8f9d6d8e62970540f427c486e/twx/botapi/botapi.py#L4422-L4424 |
4,566 | horejsek/python-webdriverwrapper | webdriverwrapper/wrapper.py | _WebdriverWrapper.html | def html(self):
"""
Returns ``innerHTML`` of whole page. On page have to be tag ``body``.
.. versionadded:: 2.2
"""
try:
body = self.get_elm(tag_name='body')
except selenium_exc.NoSuchElementException:
return None
else:
return ... | python | def html(self):
"""
Returns ``innerHTML`` of whole page. On page have to be tag ``body``.
.. versionadded:: 2.2
"""
try:
body = self.get_elm(tag_name='body')
except selenium_exc.NoSuchElementException:
return None
else:
return ... | ['def', 'html', '(', 'self', ')', ':', 'try', ':', 'body', '=', 'self', '.', 'get_elm', '(', 'tag_name', '=', "'body'", ')', 'except', 'selenium_exc', '.', 'NoSuchElementException', ':', 'return', 'None', 'else', ':', 'return', 'body', '.', 'get_attribute', '(', "'innerHTML'", ')'] | Returns ``innerHTML`` of whole page. On page have to be tag ``body``.
.. versionadded:: 2.2 | ['Returns', 'innerHTML', 'of', 'whole', 'page', '.', 'On', 'page', 'have', 'to', 'be', 'tag', 'body', '.'] | train | https://github.com/horejsek/python-webdriverwrapper/blob/a492f79ab60ed83d860dd817b6a0961500d7e3f5/webdriverwrapper/wrapper.py#L408-L419 |
4,567 | astraw38/lint | lint/validators/pylint_validator.py | no_new_errors | def no_new_errors(new_data, old_data, strict=False):
"""
Pylint Validator that will fail any review if there are
new Pylint errors in it (Pylint message starts with 'E:')
:param new_data:
:param old_data:
:return:
"""
success = True
score = 0
message = ''
if new_data['errors... | python | def no_new_errors(new_data, old_data, strict=False):
"""
Pylint Validator that will fail any review if there are
new Pylint errors in it (Pylint message starts with 'E:')
:param new_data:
:param old_data:
:return:
"""
success = True
score = 0
message = ''
if new_data['errors... | ['def', 'no_new_errors', '(', 'new_data', ',', 'old_data', ',', 'strict', '=', 'False', ')', ':', 'success', '=', 'True', 'score', '=', '0', 'message', '=', "''", 'if', 'new_data', '[', "'errors'", ']', '>', 'old_data', '[', "'errors'", ']', ':', 'success', '=', 'False', 'message', '=', '"Failed, More errors than prior... | Pylint Validator that will fail any review if there are
new Pylint errors in it (Pylint message starts with 'E:')
:param new_data:
:param old_data:
:return: | ['Pylint', 'Validator', 'that', 'will', 'fail', 'any', 'review', 'if', 'there', 'are', 'new', 'Pylint', 'errors', 'in', 'it', '(', 'Pylint', 'message', 'starts', 'with', 'E', ':', ')'] | train | https://github.com/astraw38/lint/blob/162ceefcb812f07d18544aaa887b9ec4f102cfb1/lint/validators/pylint_validator.py#L49-L69 |
4,568 | StackStorm/pybind | pybind/nos/v7_2_0/rbridge_id/ipv6/static_ag_ipv6_config/__init__.py | static_ag_ipv6_config._set_anycast_gateway_mac | def _set_anycast_gateway_mac(self, v, load=False):
"""
Setter method for anycast_gateway_mac, mapped from YANG variable /rbridge_id/ipv6/static_ag_ipv6_config/anycast_gateway_mac (container)
If this variable is read-only (config: false) in the
source YANG file, then _set_anycast_gateway_mac is considere... | python | def _set_anycast_gateway_mac(self, v, load=False):
"""
Setter method for anycast_gateway_mac, mapped from YANG variable /rbridge_id/ipv6/static_ag_ipv6_config/anycast_gateway_mac (container)
If this variable is read-only (config: false) in the
source YANG file, then _set_anycast_gateway_mac is considere... | ['def', '_set_anycast_gateway_mac', '(', 'self', ',', 'v', ',', 'load', '=', 'False', ')', ':', 'if', 'hasattr', '(', 'v', ',', '"_utype"', ')', ':', 'v', '=', 'v', '.', '_utype', '(', 'v', ')', 'try', ':', 't', '=', 'YANGDynClass', '(', 'v', ',', 'base', '=', 'anycast_gateway_mac', '.', 'anycast_gateway_mac', ',', 'is... | Setter method for anycast_gateway_mac, mapped from YANG variable /rbridge_id/ipv6/static_ag_ipv6_config/anycast_gateway_mac (container)
If this variable is read-only (config: false) in the
source YANG file, then _set_anycast_gateway_mac is considered as a private
method. Backends looking to populate this va... | ['Setter', 'method', 'for', 'anycast_gateway_mac', 'mapped', 'from', 'YANG', 'variable', '/', 'rbridge_id', '/', 'ipv6', '/', 'static_ag_ipv6_config', '/', 'anycast_gateway_mac', '(', 'container', ')', 'If', 'this', 'variable', 'is', 'read', '-', 'only', '(', 'config', ':', 'false', ')', 'in', 'the', 'source', 'YANG', ... | train | https://github.com/StackStorm/pybind/blob/44c467e71b2b425be63867aba6e6fa28b2cfe7fb/pybind/nos/v7_2_0/rbridge_id/ipv6/static_ag_ipv6_config/__init__.py#L94-L117 |
4,569 | kubernetes-client/python | kubernetes/client/api_client.py | ApiClient.update_params_for_auth | def update_params_for_auth(self, headers, querys, auth_settings):
"""
Updates header and query params based on authentication setting.
:param headers: Header parameters dict to be updated.
:param querys: Query parameters tuple list to be updated.
:param auth_settings: Authentica... | python | def update_params_for_auth(self, headers, querys, auth_settings):
"""
Updates header and query params based on authentication setting.
:param headers: Header parameters dict to be updated.
:param querys: Query parameters tuple list to be updated.
:param auth_settings: Authentica... | ['def', 'update_params_for_auth', '(', 'self', ',', 'headers', ',', 'querys', ',', 'auth_settings', ')', ':', 'if', 'not', 'auth_settings', ':', 'return', 'for', 'auth', 'in', 'auth_settings', ':', 'auth_setting', '=', 'self', '.', 'configuration', '.', 'auth_settings', '(', ')', '.', 'get', '(', 'auth', ')', 'if', 'au... | Updates header and query params based on authentication setting.
:param headers: Header parameters dict to be updated.
:param querys: Query parameters tuple list to be updated.
:param auth_settings: Authentication setting identifiers list. | ['Updates', 'header', 'and', 'query', 'params', 'based', 'on', 'authentication', 'setting', '.'] | train | https://github.com/kubernetes-client/python/blob/5e512ff564c244c50cab780d821542ed56aa965a/kubernetes/client/api_client.py#L500-L523 |
4,570 | fermiPy/fermipy | fermipy/diffuse/residual_cr.py | ResidualCR._compute_counts_from_intensity | def _compute_counts_from_intensity(intensity, bexpcube):
""" Make the counts map from the intensity
"""
data = intensity.data * np.sqrt(bexpcube.data[1:] * bexpcube.data[0:-1])
return HpxMap(data, intensity.hpx) | python | def _compute_counts_from_intensity(intensity, bexpcube):
""" Make the counts map from the intensity
"""
data = intensity.data * np.sqrt(bexpcube.data[1:] * bexpcube.data[0:-1])
return HpxMap(data, intensity.hpx) | ['def', '_compute_counts_from_intensity', '(', 'intensity', ',', 'bexpcube', ')', ':', 'data', '=', 'intensity', '.', 'data', '*', 'np', '.', 'sqrt', '(', 'bexpcube', '.', 'data', '[', '1', ':', ']', '*', 'bexpcube', '.', 'data', '[', '0', ':', '-', '1', ']', ')', 'return', 'HpxMap', '(', 'data', ',', 'intensity', '.',... | Make the counts map from the intensity | ['Make', 'the', 'counts', 'map', 'from', 'the', 'intensity'] | train | https://github.com/fermiPy/fermipy/blob/9df5e7e3728307fd58c5bba36fd86783c39fbad4/fermipy/diffuse/residual_cr.py#L142-L146 |
4,571 | saltstack/salt | salt/utils/process.py | get_pidfile | def get_pidfile(pidfile):
'''
Return the pid from a pidfile as an integer
'''
try:
with salt.utils.files.fopen(pidfile) as pdf:
pid = pdf.read().strip()
return int(pid)
except (OSError, IOError, TypeError, ValueError):
return -1 | python | def get_pidfile(pidfile):
'''
Return the pid from a pidfile as an integer
'''
try:
with salt.utils.files.fopen(pidfile) as pdf:
pid = pdf.read().strip()
return int(pid)
except (OSError, IOError, TypeError, ValueError):
return -1 | ['def', 'get_pidfile', '(', 'pidfile', ')', ':', 'try', ':', 'with', 'salt', '.', 'utils', '.', 'files', '.', 'fopen', '(', 'pidfile', ')', 'as', 'pdf', ':', 'pid', '=', 'pdf', '.', 'read', '(', ')', '.', 'strip', '(', ')', 'return', 'int', '(', 'pid', ')', 'except', '(', 'OSError', ',', 'IOError', ',', 'TypeError', ',... | Return the pid from a pidfile as an integer | ['Return', 'the', 'pid', 'from', 'a', 'pidfile', 'as', 'an', 'integer'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/utils/process.py#L225-L234 |
4,572 | mikeywaites/flask-arrested | arrested/mixins.py | ObjectMixin.obj | def obj(self):
"""Returns the value of :meth:`ObjectMixin.get_object` and sets a private
property called _obj. This property ensures the logic around allow_none
is enforced across Endpoints using the Object interface.
:raises: :class:`werkzeug.exceptions.BadRequest`
:returns: T... | python | def obj(self):
"""Returns the value of :meth:`ObjectMixin.get_object` and sets a private
property called _obj. This property ensures the logic around allow_none
is enforced across Endpoints using the Object interface.
:raises: :class:`werkzeug.exceptions.BadRequest`
:returns: T... | ['def', 'obj', '(', 'self', ')', ':', 'if', 'not', 'getattr', '(', 'self', ',', "'_obj'", ',', 'None', ')', ':', 'self', '.', '_obj', '=', 'self', '.', 'get_object', '(', ')', 'if', 'self', '.', '_obj', 'is', 'None', 'and', 'not', 'self', '.', 'allow_none', ':', 'self', '.', 'return_error', '(', '404', ')', 'return', '... | Returns the value of :meth:`ObjectMixin.get_object` and sets a private
property called _obj. This property ensures the logic around allow_none
is enforced across Endpoints using the Object interface.
:raises: :class:`werkzeug.exceptions.BadRequest`
:returns: The result of :meth:ObjectM... | ['Returns', 'the', 'value', 'of', ':', 'meth', ':', 'ObjectMixin', '.', 'get_object', 'and', 'sets', 'a', 'private', 'property', 'called', '_obj', '.', 'This', 'property', 'ensures', 'the', 'logic', 'around', 'allow_none', 'is', 'enforced', 'across', 'Endpoints', 'using', 'the', 'Object', 'interface', '.'] | train | https://github.com/mikeywaites/flask-arrested/blob/6b97ce2ad2765f9acab10f4726e310258aa51de0/arrested/mixins.py#L122-L135 |
4,573 | lambdamusic/Ontospy | ontospy/core/utils.py | try_sort_fmt_opts | def try_sort_fmt_opts(rdf_format_opts_list, uri):
"""reorder fmt options based on uri file type suffix - if available - so to test most likely serialization first when parsing some RDF
NOTE this is not very nice as it is hardcoded and assumes the origin serializations to be this: ['turtle', 'xml', 'n3', 'nt',... | python | def try_sort_fmt_opts(rdf_format_opts_list, uri):
"""reorder fmt options based on uri file type suffix - if available - so to test most likely serialization first when parsing some RDF
NOTE this is not very nice as it is hardcoded and assumes the origin serializations to be this: ['turtle', 'xml', 'n3', 'nt',... | ['def', 'try_sort_fmt_opts', '(', 'rdf_format_opts_list', ',', 'uri', ')', ':', 'filename', ',', 'file_extension', '=', 'os', '.', 'path', '.', 'splitext', '(', 'uri', ')', '# print(filename, file_extension)', 'if', 'file_extension', '==', '".ttl"', 'or', 'file_extension', '==', '".turtle"', ':', 'return', '[', "'turtl... | reorder fmt options based on uri file type suffix - if available - so to test most likely serialization first when parsing some RDF
NOTE this is not very nice as it is hardcoded and assumes the origin serializations to be this: ['turtle', 'xml', 'n3', 'nt', 'json-ld', 'rdfa'] | ['reorder', 'fmt', 'options', 'based', 'on', 'uri', 'file', 'type', 'suffix', '-', 'if', 'available', '-', 'so', 'to', 'test', 'most', 'likely', 'serialization', 'first', 'when', 'parsing', 'some', 'RDF'] | train | https://github.com/lambdamusic/Ontospy/blob/eb46cb13792b2b87f21babdf976996318eec7571/ontospy/core/utils.py#L893-L926 |
4,574 | hollenstein/maspy | maspy/reader.py | applySiiRanking | def applySiiRanking(siiContainer, specfile):
"""Iterates over all Sii entries of a specfile in siiContainer and sorts Sii
elements of the same spectrum according to the score attribute specified in
``siiContainer.info[specfile]['rankAttr']``. Sorted Sii elements are then
ranked according to their sorte... | python | def applySiiRanking(siiContainer, specfile):
"""Iterates over all Sii entries of a specfile in siiContainer and sorts Sii
elements of the same spectrum according to the score attribute specified in
``siiContainer.info[specfile]['rankAttr']``. Sorted Sii elements are then
ranked according to their sorte... | ['def', 'applySiiRanking', '(', 'siiContainer', ',', 'specfile', ')', ':', 'attr', '=', 'siiContainer', '.', 'info', '[', 'specfile', ']', '[', "'rankAttr'", ']', 'reverse', '=', 'siiContainer', '.', 'info', '[', 'specfile', ']', '[', "'rankLargerBetter'", ']', 'for', 'itemList', 'in', 'listvalues', '(', 'siiContainer'... | Iterates over all Sii entries of a specfile in siiContainer and sorts Sii
elements of the same spectrum according to the score attribute specified in
``siiContainer.info[specfile]['rankAttr']``. Sorted Sii elements are then
ranked according to their sorted position, if multiple Sii have the same
score,... | ['Iterates', 'over', 'all', 'Sii', 'entries', 'of', 'a', 'specfile', 'in', 'siiContainer', 'and', 'sorts', 'Sii', 'elements', 'of', 'the', 'same', 'spectrum', 'according', 'to', 'the', 'score', 'attribute', 'specified', 'in', 'siiContainer', '.', 'info', '[', 'specfile', ']', '[', 'rankAttr', ']', '.', 'Sorted', 'Sii',... | train | https://github.com/hollenstein/maspy/blob/f15fcfd24df306d8420540460d902aa3073ec133/maspy/reader.py#L336-L359 |
4,575 | RPi-Distro/python-sense-hat | sense_hat/sense_hat.py | SenseHat._unpack_bin | def _unpack_bin(self, packed):
"""
Internal. Decodes 16 bit RGB565 into python list [R,G,B]
"""
output = struct.unpack('H', packed)
bits16 = output[0]
r = (bits16 & 0xF800) >> 11
g = (bits16 & 0x7E0) >> 5
b = (bits16 & 0x1F)
return [int(r << 3), i... | python | def _unpack_bin(self, packed):
"""
Internal. Decodes 16 bit RGB565 into python list [R,G,B]
"""
output = struct.unpack('H', packed)
bits16 = output[0]
r = (bits16 & 0xF800) >> 11
g = (bits16 & 0x7E0) >> 5
b = (bits16 & 0x1F)
return [int(r << 3), i... | ['def', '_unpack_bin', '(', 'self', ',', 'packed', ')', ':', 'output', '=', 'struct', '.', 'unpack', '(', "'H'", ',', 'packed', ')', 'bits16', '=', 'output', '[', '0', ']', 'r', '=', '(', 'bits16', '&', '0xF800', ')', '>>', '11', 'g', '=', '(', 'bits16', '&', '0x7E0', ')', '>>', '5', 'b', '=', '(', 'bits16', '&', '0x1F... | Internal. Decodes 16 bit RGB565 into python list [R,G,B] | ['Internal', '.', 'Decodes', '16', 'bit', 'RGB565', 'into', 'python', 'list', '[', 'R', 'G', 'B', ']'] | train | https://github.com/RPi-Distro/python-sense-hat/blob/9a37f0923ce8dbde69514c3b8d58d30de01c9ee7/sense_hat/sense_hat.py#L232-L242 |
4,576 | Fizzadar/pyinfra | pyinfra/modules/postgresql.py | database | def database(
state, host, name,
present=True, owner=None,
template=None, encoding=None,
lc_collate=None, lc_ctype=None, tablespace=None,
connection_limit=None,
# Details for speaking to PostgreSQL via `psql` CLI
postgresql_user=None, postgresql_password=None,
postgresql_host=None, postg... | python | def database(
state, host, name,
present=True, owner=None,
template=None, encoding=None,
lc_collate=None, lc_ctype=None, tablespace=None,
connection_limit=None,
# Details for speaking to PostgreSQL via `psql` CLI
postgresql_user=None, postgresql_password=None,
postgresql_host=None, postg... | ['def', 'database', '(', 'state', ',', 'host', ',', 'name', ',', 'present', '=', 'True', ',', 'owner', '=', 'None', ',', 'template', '=', 'None', ',', 'encoding', '=', 'None', ',', 'lc_collate', '=', 'None', ',', 'lc_ctype', '=', 'None', ',', 'tablespace', '=', 'None', ',', 'connection_limit', '=', 'None', ',', '# Deta... | Add/remove PostgreSQL databases.
+ name: name of the database
+ present: whether the database should exist or not
+ owner: the PostgreSQL role that owns the database
+ template: name of the PostgreSQL template to use
+ encoding: encoding of the database
+ lc_collate: lc_collate of the database
... | ['Add', '/', 'remove', 'PostgreSQL', 'databases', '.'] | train | https://github.com/Fizzadar/pyinfra/blob/006f751f7db2e07d32522c0285160783de2feb79/pyinfra/modules/postgresql.py#L127-L197 |
4,577 | skyfielders/python-skyfield | skyfield/units.py | Distance.to | def to(self, unit):
"""Convert this distance to the given AstroPy unit."""
from astropy.units import au
return (self.au * au).to(unit) | python | def to(self, unit):
"""Convert this distance to the given AstroPy unit."""
from astropy.units import au
return (self.au * au).to(unit) | ['def', 'to', '(', 'self', ',', 'unit', ')', ':', 'from', 'astropy', '.', 'units', 'import', 'au', 'return', '(', 'self', '.', 'au', '*', 'au', ')', '.', 'to', '(', 'unit', ')'] | Convert this distance to the given AstroPy unit. | ['Convert', 'this', 'distance', 'to', 'the', 'given', 'AstroPy', 'unit', '.'] | train | https://github.com/skyfielders/python-skyfield/blob/51d9e042e06457f6b1f2415296d50a38cb3a300f/skyfield/units.py#L76-L79 |
4,578 | yangl1996/libpagure | libpagure/libpagure.py | Pagure.project_branches | def project_branches(self):
"""
List all branches associated with a repository.
:return:
"""
request_url = "{}git/branches".format(self.create_basic_url())
return_value = self._call_api(request_url)
return return_value['branches'] | python | def project_branches(self):
"""
List all branches associated with a repository.
:return:
"""
request_url = "{}git/branches".format(self.create_basic_url())
return_value = self._call_api(request_url)
return return_value['branches'] | ['def', 'project_branches', '(', 'self', ')', ':', 'request_url', '=', '"{}git/branches"', '.', 'format', '(', 'self', '.', 'create_basic_url', '(', ')', ')', 'return_value', '=', 'self', '.', '_call_api', '(', 'request_url', ')', 'return', 'return_value', '[', "'branches'", ']'] | List all branches associated with a repository.
:return: | ['List', 'all', 'branches', 'associated', 'with', 'a', 'repository', '.', ':', 'return', ':'] | train | https://github.com/yangl1996/libpagure/blob/dd96ed29142407463790c66ed321984a6ea7465a/libpagure/libpagure.py#L661-L670 |
4,579 | amoffat/sh | sh.py | ob_is_tty | def ob_is_tty(ob):
""" checks if an object (like a file-like object) is a tty. """
fileno = get_fileno(ob)
is_tty = False
if fileno:
is_tty = os.isatty(fileno)
return is_tty | python | def ob_is_tty(ob):
""" checks if an object (like a file-like object) is a tty. """
fileno = get_fileno(ob)
is_tty = False
if fileno:
is_tty = os.isatty(fileno)
return is_tty | ['def', 'ob_is_tty', '(', 'ob', ')', ':', 'fileno', '=', 'get_fileno', '(', 'ob', ')', 'is_tty', '=', 'False', 'if', 'fileno', ':', 'is_tty', '=', 'os', '.', 'isatty', '(', 'fileno', ')', 'return', 'is_tty'] | checks if an object (like a file-like object) is a tty. | ['checks', 'if', 'an', 'object', '(', 'like', 'a', 'file', '-', 'like', 'object', ')', 'is', 'a', 'tty', '.'] | train | https://github.com/amoffat/sh/blob/858adf0c682af4c40e41f34d6926696b7a5d3b12/sh.py#L997-L1003 |
4,580 | rackerlabs/simpl | simpl/config.py | Option.add_argument | def add_argument(self, parser, permissive=False, **override_kwargs):
"""Add an option to a an argparse parser.
:keyword permissive: when true, build a parser that does not validate
required arguments.
"""
kwargs = {}
required = None
if self.kwargs:
... | python | def add_argument(self, parser, permissive=False, **override_kwargs):
"""Add an option to a an argparse parser.
:keyword permissive: when true, build a parser that does not validate
required arguments.
"""
kwargs = {}
required = None
if self.kwargs:
... | ['def', 'add_argument', '(', 'self', ',', 'parser', ',', 'permissive', '=', 'False', ',', '*', '*', 'override_kwargs', ')', ':', 'kwargs', '=', '{', '}', 'required', '=', 'None', 'if', 'self', '.', 'kwargs', ':', 'kwargs', '=', 'copy', '.', 'copy', '(', 'self', '.', 'kwargs', ')', 'if', "'env'", 'in', 'kwargs', 'and', ... | Add an option to a an argparse parser.
:keyword permissive: when true, build a parser that does not validate
required arguments. | ['Add', 'an', 'option', 'to', 'a', 'an', 'argparse', 'parser', '.'] | train | https://github.com/rackerlabs/simpl/blob/60ed3336a931cd6a7a7246e60f26165d9dc7c99c/simpl/config.py#L248-L311 |
4,581 | jonhadfield/python-hosts | python_hosts/hosts.py | HostsEntry.get_entry_type | def get_entry_type(hosts_entry=None):
"""
Return the type of entry for the line of hosts file passed
:param hosts_entry: A line from the hosts file
:return: 'comment' | 'blank' | 'ipv4' | 'ipv6'
"""
if hosts_entry and isinstance(hosts_entry, str):
entry = host... | python | def get_entry_type(hosts_entry=None):
"""
Return the type of entry for the line of hosts file passed
:param hosts_entry: A line from the hosts file
:return: 'comment' | 'blank' | 'ipv4' | 'ipv6'
"""
if hosts_entry and isinstance(hosts_entry, str):
entry = host... | ['def', 'get_entry_type', '(', 'hosts_entry', '=', 'None', ')', ':', 'if', 'hosts_entry', 'and', 'isinstance', '(', 'hosts_entry', ',', 'str', ')', ':', 'entry', '=', 'hosts_entry', '.', 'strip', '(', ')', 'if', 'not', 'entry', 'or', 'not', 'entry', '[', '0', ']', 'or', 'entry', '[', '0', ']', '==', '"\\n"', ':', 'retu... | Return the type of entry for the line of hosts file passed
:param hosts_entry: A line from the hosts file
:return: 'comment' | 'blank' | 'ipv4' | 'ipv6' | ['Return', 'the', 'type', 'of', 'entry', 'for', 'the', 'line', 'of', 'hosts', 'file', 'passed', ':', 'param', 'hosts_entry', ':', 'A', 'line', 'from', 'the', 'hosts', 'file', ':', 'return', ':', 'comment', '|', 'blank', '|', 'ipv4', '|', 'ipv6'] | train | https://github.com/jonhadfield/python-hosts/blob/9ccaa8edc63418a91f10bf732b26070f21dd2ad0/python_hosts/hosts.py#L90-L106 |
4,582 | readbeyond/aeneas | aeneas/tree.py | Tree.level_at_index | def level_at_index(self, index):
"""
Return the list of nodes at level ``index``,
in DFS order.
:param int index: the index
:rtype: list of :class:`~aeneas.tree.Tree`
:raises: ValueError if the given ``index`` is not valid
"""
if not isinstance(index, in... | python | def level_at_index(self, index):
"""
Return the list of nodes at level ``index``,
in DFS order.
:param int index: the index
:rtype: list of :class:`~aeneas.tree.Tree`
:raises: ValueError if the given ``index`` is not valid
"""
if not isinstance(index, in... | ['def', 'level_at_index', '(', 'self', ',', 'index', ')', ':', 'if', 'not', 'isinstance', '(', 'index', ',', 'int', ')', ':', 'self', '.', 'log_exc', '(', 'u"Index is not an integer"', ',', 'None', ',', 'True', ',', 'TypeError', ')', 'levels', '=', 'self', '.', 'levels', 'if', '(', 'index', '<', '0', ')', 'or', '(', 'i... | Return the list of nodes at level ``index``,
in DFS order.
:param int index: the index
:rtype: list of :class:`~aeneas.tree.Tree`
:raises: ValueError if the given ``index`` is not valid | ['Return', 'the', 'list', 'of', 'nodes', 'at', 'level', 'index', 'in', 'DFS', 'order', '.'] | train | https://github.com/readbeyond/aeneas/blob/9d95535ad63eef4a98530cfdff033b8c35315ee1/aeneas/tree.py#L425-L440 |
4,583 | glut23/webvtt-py | webvtt/parsers.py | TextBasedParser._parse_timeframe_line | def _parse_timeframe_line(self, line):
"""Parse timeframe line and return start and end timestamps."""
tf = self._validate_timeframe_line(line)
if not tf:
raise MalformedCaptionError('Invalid time format')
return tf.group(1), tf.group(2) | python | def _parse_timeframe_line(self, line):
"""Parse timeframe line and return start and end timestamps."""
tf = self._validate_timeframe_line(line)
if not tf:
raise MalformedCaptionError('Invalid time format')
return tf.group(1), tf.group(2) | ['def', '_parse_timeframe_line', '(', 'self', ',', 'line', ')', ':', 'tf', '=', 'self', '.', '_validate_timeframe_line', '(', 'line', ')', 'if', 'not', 'tf', ':', 'raise', 'MalformedCaptionError', '(', "'Invalid time format'", ')', 'return', 'tf', '.', 'group', '(', '1', ')', ',', 'tf', '.', 'group', '(', '2', ')'] | Parse timeframe line and return start and end timestamps. | ['Parse', 'timeframe', 'line', 'and', 'return', 'start', 'and', 'end', 'timestamps', '.'] | train | https://github.com/glut23/webvtt-py/blob/7b4da0123c2e2afaf31402107528721eb1d3d481/webvtt/parsers.py#L49-L55 |
4,584 | linuxsoftware/ls.joyous | ls/joyous/models/events.py | RecurringEventPage.status | def status(self):
"""
The current status of the event (started, finished or pending).
"""
myNow = timezone.localtime(timezone=self.tz)
daysDelta = dt.timedelta(days=self.num_days - 1)
# NB: postponements can be created after the until date
# so ignore that
... | python | def status(self):
"""
The current status of the event (started, finished or pending).
"""
myNow = timezone.localtime(timezone=self.tz)
daysDelta = dt.timedelta(days=self.num_days - 1)
# NB: postponements can be created after the until date
# so ignore that
... | ['def', 'status', '(', 'self', ')', ':', 'myNow', '=', 'timezone', '.', 'localtime', '(', 'timezone', '=', 'self', '.', 'tz', ')', 'daysDelta', '=', 'dt', '.', 'timedelta', '(', 'days', '=', 'self', '.', 'num_days', '-', '1', ')', '# NB: postponements can be created after the until date', '# so ignore that', 'today... | The current status of the event (started, finished or pending). | ['The', 'current', 'status', 'of', 'the', 'event', '(', 'started', 'finished', 'or', 'pending', ')', '.'] | train | https://github.com/linuxsoftware/ls.joyous/blob/316283140ca5171a68ad3170a5964fdc89be0b56/ls/joyous/models/events.py#L1003-L1026 |
4,585 | benoitkugler/abstractDataLibrary | pyDLib/Core/formats.py | abstractSearch.in_date | def in_date(objet, pattern):
""" abstractSearch dans une date datetime.date"""
if objet:
pattern = re.sub(" ", '', pattern)
objet_str = abstractRender.date(objet)
return bool(re.search(pattern, objet_str))
return False | python | def in_date(objet, pattern):
""" abstractSearch dans une date datetime.date"""
if objet:
pattern = re.sub(" ", '', pattern)
objet_str = abstractRender.date(objet)
return bool(re.search(pattern, objet_str))
return False | ['def', 'in_date', '(', 'objet', ',', 'pattern', ')', ':', 'if', 'objet', ':', 'pattern', '=', 're', '.', 'sub', '(', '" "', ',', "''", ',', 'pattern', ')', 'objet_str', '=', 'abstractRender', '.', 'date', '(', 'objet', ')', 'return', 'bool', '(', 're', '.', 'search', '(', 'pattern', ',', 'objet_str', ')', ')', 'return... | abstractSearch dans une date datetime.date | ['abstractSearch', 'dans', 'une', 'date', 'datetime', '.', 'date'] | train | https://github.com/benoitkugler/abstractDataLibrary/blob/16be28e99837e40287a63803bbfdf67ac1806b7b/pyDLib/Core/formats.py#L204-L210 |
4,586 | libChEBI/libChEBIpy | libchebipy/_parsers.py | get_mol_filename | def get_mol_filename(chebi_id):
'''Returns mol file'''
mol = get_mol(chebi_id)
if mol is None:
return None
file_descriptor, mol_filename = tempfile.mkstemp(str(chebi_id) +
'_', '.mol')
mol_file = open(mol_filename, 'w')
mol_file.writ... | python | def get_mol_filename(chebi_id):
'''Returns mol file'''
mol = get_mol(chebi_id)
if mol is None:
return None
file_descriptor, mol_filename = tempfile.mkstemp(str(chebi_id) +
'_', '.mol')
mol_file = open(mol_filename, 'w')
mol_file.writ... | ['def', 'get_mol_filename', '(', 'chebi_id', ')', ':', 'mol', '=', 'get_mol', '(', 'chebi_id', ')', 'if', 'mol', 'is', 'None', ':', 'return', 'None', 'file_descriptor', ',', 'mol_filename', '=', 'tempfile', '.', 'mkstemp', '(', 'str', '(', 'chebi_id', ')', '+', "'_'", ',', "'.mol'", ')', 'mol_file', '=', 'open', '(', '... | Returns mol file | ['Returns', 'mol', 'file'] | train | https://github.com/libChEBI/libChEBIpy/blob/89f223a91f518619d5e3910070d283adcac1626e/libchebipy/_parsers.py#L570-L584 |
4,587 | pyeve/cerberus | cerberus/validator.py | BareValidator.__validate_definitions | def __validate_definitions(self, definitions, field):
""" Validate a field's value against its defined rules. """
def validate_rule(rule):
validator = self.__get_rule_handler('validate', rule)
return validator(definitions.get(rule, None), field, value)
definitions = sel... | python | def __validate_definitions(self, definitions, field):
""" Validate a field's value against its defined rules. """
def validate_rule(rule):
validator = self.__get_rule_handler('validate', rule)
return validator(definitions.get(rule, None), field, value)
definitions = sel... | ['def', '__validate_definitions', '(', 'self', ',', 'definitions', ',', 'field', ')', ':', 'def', 'validate_rule', '(', 'rule', ')', ':', 'validator', '=', 'self', '.', '__get_rule_handler', '(', "'validate'", ',', 'rule', ')', 'return', 'validator', '(', 'definitions', '.', 'get', '(', 'rule', ',', 'None', ')', ',', '... | Validate a field's value against its defined rules. | ['Validate', 'a', 'field', 's', 'value', 'against', 'its', 'defined', 'rules', '.'] | train | https://github.com/pyeve/cerberus/blob/688a67a4069e88042ed424bda7be0f4fa5fc3910/cerberus/validator.py#L1036-L1073 |
4,588 | iotile/coretools | iotileship/iotile/ship/autobuild/ship_file.py | autobuild_shiparchive | def autobuild_shiparchive(src_file):
"""Create a ship file archive containing a yaml_file and its dependencies.
If yaml_file depends on any build products as external files, it must
be a jinja2 template that references the file using the find_product
filter so that we can figure out where those build p... | python | def autobuild_shiparchive(src_file):
"""Create a ship file archive containing a yaml_file and its dependencies.
If yaml_file depends on any build products as external files, it must
be a jinja2 template that references the file using the find_product
filter so that we can figure out where those build p... | ['def', 'autobuild_shiparchive', '(', 'src_file', ')', ':', 'if', 'not', 'src_file', '.', 'endswith', '(', "'.tpl'", ')', ':', 'raise', 'BuildError', '(', '"You must pass a .tpl file to autobuild_shiparchive"', ',', 'src_file', '=', 'src_file', ')', 'env', '=', 'Environment', '(', 'tools', '=', '[', ']', ')', 'family',... | Create a ship file archive containing a yaml_file and its dependencies.
If yaml_file depends on any build products as external files, it must
be a jinja2 template that references the file using the find_product
filter so that we can figure out where those build products are going
and create the right d... | ['Create', 'a', 'ship', 'file', 'archive', 'containing', 'a', 'yaml_file', 'and', 'its', 'dependencies', '.'] | train | https://github.com/iotile/coretools/blob/2d794f5f1346b841b0dcd16c9d284e9bf2f3c6ec/iotileship/iotile/ship/autobuild/ship_file.py#L15-L88 |
4,589 | richardkiss/pycoin | pycoin/crack/bip32.py | ascend_bip32 | def ascend_bip32(bip32_pub_node, secret_exponent, child):
"""
Given a BIP32Node with public derivation child "child" with a known private key,
return the secret exponent for the bip32_pub_node.
"""
i_as_bytes = struct.pack(">l", child)
sec = public_pair_to_sec(bip32_pub_node.public_pair(), compr... | python | def ascend_bip32(bip32_pub_node, secret_exponent, child):
"""
Given a BIP32Node with public derivation child "child" with a known private key,
return the secret exponent for the bip32_pub_node.
"""
i_as_bytes = struct.pack(">l", child)
sec = public_pair_to_sec(bip32_pub_node.public_pair(), compr... | ['def', 'ascend_bip32', '(', 'bip32_pub_node', ',', 'secret_exponent', ',', 'child', ')', ':', 'i_as_bytes', '=', 'struct', '.', 'pack', '(', '">l"', ',', 'child', ')', 'sec', '=', 'public_pair_to_sec', '(', 'bip32_pub_node', '.', 'public_pair', '(', ')', ',', 'compressed', '=', 'True', ')', 'data', '=', 'sec', '+', 'i... | Given a BIP32Node with public derivation child "child" with a known private key,
return the secret exponent for the bip32_pub_node. | ['Given', 'a', 'BIP32Node', 'with', 'public', 'derivation', 'child', 'child', 'with', 'a', 'known', 'private', 'key', 'return', 'the', 'secret', 'exponent', 'for', 'the', 'bip32_pub_node', '.'] | train | https://github.com/richardkiss/pycoin/blob/1e8d0d9fe20ce0347b97847bb529cd1bd84c7442/pycoin/crack/bip32.py#L9-L19 |
4,590 | DataONEorg/d1_python | lib_common/src/d1_common/types/scripts/pyxbgen_all.py | GenerateBindings.run_pyxbgen | def run_pyxbgen(self, args):
"""Args:
args:
"""
cmd = 'pyxbgen {}'.format(' '.join(args))
print(cmd)
os.system(cmd) | python | def run_pyxbgen(self, args):
"""Args:
args:
"""
cmd = 'pyxbgen {}'.format(' '.join(args))
print(cmd)
os.system(cmd) | ['def', 'run_pyxbgen', '(', 'self', ',', 'args', ')', ':', 'cmd', '=', "'pyxbgen {}'", '.', 'format', '(', "' '", '.', 'join', '(', 'args', ')', ')', 'print', '(', 'cmd', ')', 'os', '.', 'system', '(', 'cmd', ')'] | Args:
args: | ['Args', ':'] | train | https://github.com/DataONEorg/d1_python/blob/3ac4d4f3ca052d3e8641a6a329cab526c8ddcb0d/lib_common/src/d1_common/types/scripts/pyxbgen_all.py#L149-L157 |
4,591 | cuihantao/andes | andes/models/base.py | ModelBase.link_bus | def link_bus(self, bus_idx):
"""
Return the indices of elements linking the given buses
:param bus_idx:
:return:
"""
ret = []
if not self._config['is_series']:
self.log(
'link_bus function is not valid for non-series model <{}>'.
... | python | def link_bus(self, bus_idx):
"""
Return the indices of elements linking the given buses
:param bus_idx:
:return:
"""
ret = []
if not self._config['is_series']:
self.log(
'link_bus function is not valid for non-series model <{}>'.
... | ['def', 'link_bus', '(', 'self', ',', 'bus_idx', ')', ':', 'ret', '=', '[', ']', 'if', 'not', 'self', '.', '_config', '[', "'is_series'", ']', ':', 'self', '.', 'log', '(', "'link_bus function is not valid for non-series model <{}>'", '.', 'format', '(', 'self', '.', 'name', ')', ')', 'return', '[', ']', 'if', 'isinsta... | Return the indices of elements linking the given buses
:param bus_idx:
:return: | ['Return', 'the', 'indices', 'of', 'elements', 'linking', 'the', 'given', 'buses'] | train | https://github.com/cuihantao/andes/blob/7067898d4f26ce7534e968b8486c4aa8fe3a511a/andes/models/base.py#L1398-L1441 |
4,592 | saltstack/salt | salt/states/azurearm_compute.py | availability_set_present | def availability_set_present(name, resource_group, tags=None, platform_update_domain_count=None,
platform_fault_domain_count=None, virtual_machines=None, sku=None, connection_auth=None,
**kwargs):
'''
.. versionadded:: 2019.2.0
Ensure an availabilit... | python | def availability_set_present(name, resource_group, tags=None, platform_update_domain_count=None,
platform_fault_domain_count=None, virtual_machines=None, sku=None, connection_auth=None,
**kwargs):
'''
.. versionadded:: 2019.2.0
Ensure an availabilit... | ['def', 'availability_set_present', '(', 'name', ',', 'resource_group', ',', 'tags', '=', 'None', ',', 'platform_update_domain_count', '=', 'None', ',', 'platform_fault_domain_count', '=', 'None', ',', 'virtual_machines', '=', 'None', ',', 'sku', '=', 'None', ',', 'connection_auth', '=', 'None', ',', '*', '*', 'kwargs'... | .. versionadded:: 2019.2.0
Ensure an availability set exists.
:param name:
Name of the availability set.
:param resource_group:
The resource group assigned to the availability set.
:param tags:
A dictionary of strings can be passed as tag metadata to the availability set obje... | ['..', 'versionadded', '::', '2019', '.', '2', '.', '0'] | train | https://github.com/saltstack/salt/blob/e8541fd6e744ab0df786c0f76102e41631f45d46/salt/states/azurearm_compute.py#L104-L263 |
4,593 | iotile/coretools | iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Tool/FortranCommon.py | add_f90_to_env | def add_f90_to_env(env):
"""Add Builders and construction variables for f90 to an Environment."""
try:
F90Suffixes = env['F90FILESUFFIXES']
except KeyError:
F90Suffixes = ['.f90']
#print("Adding %s to f90 suffixes" % F90Suffixes)
try:
F90PPSuffixes = env['F90PPFILESUFFIXES']... | python | def add_f90_to_env(env):
"""Add Builders and construction variables for f90 to an Environment."""
try:
F90Suffixes = env['F90FILESUFFIXES']
except KeyError:
F90Suffixes = ['.f90']
#print("Adding %s to f90 suffixes" % F90Suffixes)
try:
F90PPSuffixes = env['F90PPFILESUFFIXES']... | ['def', 'add_f90_to_env', '(', 'env', ')', ':', 'try', ':', 'F90Suffixes', '=', 'env', '[', "'F90FILESUFFIXES'", ']', 'except', 'KeyError', ':', 'F90Suffixes', '=', '[', "'.f90'", ']', '#print("Adding %s to f90 suffixes" % F90Suffixes)', 'try', ':', 'F90PPSuffixes', '=', 'env', '[', "'F90PPFILESUFFIXES'", ']', 'except'... | Add Builders and construction variables for f90 to an Environment. | ['Add', 'Builders', 'and', 'construction', 'variables', 'for', 'f90', 'to', 'an', 'Environment', '.'] | train | https://github.com/iotile/coretools/blob/2d794f5f1346b841b0dcd16c9d284e9bf2f3c6ec/iotilebuild/iotile/build/config/scons-local-3.0.1/SCons/Tool/FortranCommon.py#L202-L216 |
4,594 | gwastro/pycbc | pycbc/dq.py | query_flag | def query_flag(ifo, name, start_time, end_time,
source='any', server="segments.ligo.org",
veto_definer=None, cache=False):
"""Return the times where the flag is active
Parameters
----------
ifo: string
The interferometer to query (H1, L1).
name: string
... | python | def query_flag(ifo, name, start_time, end_time,
source='any', server="segments.ligo.org",
veto_definer=None, cache=False):
"""Return the times where the flag is active
Parameters
----------
ifo: string
The interferometer to query (H1, L1).
name: string
... | ['def', 'query_flag', '(', 'ifo', ',', 'name', ',', 'start_time', ',', 'end_time', ',', 'source', '=', "'any'", ',', 'server', '=', '"segments.ligo.org"', ',', 'veto_definer', '=', 'None', ',', 'cache', '=', 'False', ')', ':', 'info', '=', 'name', '.', 'split', '(', "':'", ')', 'if', 'len', '(', 'info', ')', '==', '2',... | Return the times where the flag is active
Parameters
----------
ifo: string
The interferometer to query (H1, L1).
name: string
The status flag to query from LOSC.
start_time: int
The starting gps time to begin querying from LOSC
end_time: int
The end gps time of ... | ['Return', 'the', 'times', 'where', 'the', 'flag', 'is', 'active'] | train | https://github.com/gwastro/pycbc/blob/7a64cdd104d263f1b6ea0b01e6841837d05a4cb3/pycbc/dq.py#L98-L224 |
4,595 | samuel-phan/mssh-copy-id | tasks.py | build_src | def build_src(ctx, dest=None):
"""
build source archive
"""
if dest:
if not dest.startswith('/'):
# Relative
dest = os.path.join(os.getcwd(), dest)
os.chdir(PROJECT_DIR)
ctx.run('python setup.py sdist --dist-dir {0}'.format(dest))
else:
os.chd... | python | def build_src(ctx, dest=None):
"""
build source archive
"""
if dest:
if not dest.startswith('/'):
# Relative
dest = os.path.join(os.getcwd(), dest)
os.chdir(PROJECT_DIR)
ctx.run('python setup.py sdist --dist-dir {0}'.format(dest))
else:
os.chd... | ['def', 'build_src', '(', 'ctx', ',', 'dest', '=', 'None', ')', ':', 'if', 'dest', ':', 'if', 'not', 'dest', '.', 'startswith', '(', "'/'", ')', ':', '# Relative', 'dest', '=', 'os', '.', 'path', '.', 'join', '(', 'os', '.', 'getcwd', '(', ')', ',', 'dest', ')', 'os', '.', 'chdir', '(', 'PROJECT_DIR', ')', 'ctx', '.', ... | build source archive | ['build', 'source', 'archive'] | train | https://github.com/samuel-phan/mssh-copy-id/blob/59c50eabb74c4e0eeb729266df57c285e6661b0b/tasks.py#L142-L155 |
4,596 | radjkarl/imgProcessor | DUMP/interpolationMethods.py | polynomial | def polynomial(img, mask, inplace=False, replace_all=False,
max_dev=1e-5, max_iter=20, order=2):
'''
replace all masked values
calculate flatField from 2d-polynomal fit filling
all high gradient areas within averaged fit-image
returns flatField, average background level, fitte... | python | def polynomial(img, mask, inplace=False, replace_all=False,
max_dev=1e-5, max_iter=20, order=2):
'''
replace all masked values
calculate flatField from 2d-polynomal fit filling
all high gradient areas within averaged fit-image
returns flatField, average background level, fitte... | ['def', 'polynomial', '(', 'img', ',', 'mask', ',', 'inplace', '=', 'False', ',', 'replace_all', '=', 'False', ',', 'max_dev', '=', '1e-5', ',', 'max_iter', '=', '20', ',', 'order', '=', '2', ')', ':', 'if', 'inplace', ':', 'out', '=', 'img', 'else', ':', 'out', '=', 'img', '.', 'copy', '(', ')', 'lastm', '=', '0', 'fo... | replace all masked values
calculate flatField from 2d-polynomal fit filling
all high gradient areas within averaged fit-image
returns flatField, average background level, fitted image, valid indices mask | ['replace', 'all', 'masked', 'values', 'calculate', 'flatField', 'from', '2d', '-', 'polynomal', 'fit', 'filling', 'all', 'high', 'gradient', 'areas', 'within', 'averaged', 'fit', '-', 'image', 'returns', 'flatField', 'average', 'background', 'level', 'fitted', 'image', 'valid', 'indices', 'mask'] | train | https://github.com/radjkarl/imgProcessor/blob/7c5a28718f81c01a430152c60a686ac50afbfd7c/DUMP/interpolationMethods.py#L55-L88 |
4,597 | AguaClara/aguaclara | aguaclara/core/head_loss.py | _k_value_tapered_reduction | def _k_value_tapered_reduction(ent_pipe_id, exit_pipe_id, fitting_angle, re, f):
"""Returns the minor loss coefficient for a tapered reducer.
Parameters:
ent_pipe_id: Entrance pipe's inner diameter.
exit_pipe_id: Exit pipe's inner diameter.
fitting_angle: Fitting angle between entrance ... | python | def _k_value_tapered_reduction(ent_pipe_id, exit_pipe_id, fitting_angle, re, f):
"""Returns the minor loss coefficient for a tapered reducer.
Parameters:
ent_pipe_id: Entrance pipe's inner diameter.
exit_pipe_id: Exit pipe's inner diameter.
fitting_angle: Fitting angle between entrance ... | ['def', '_k_value_tapered_reduction', '(', 'ent_pipe_id', ',', 'exit_pipe_id', ',', 'fitting_angle', ',', 're', ',', 'f', ')', ':', 'k_value_square_reduction', '=', '_k_value_square_reduction', '(', 'ent_pipe_id', ',', 'exit_pipe_id', ',', 're', ',', 'f', ')', 'if', '45', '<', 'fitting_angle', '<=', '180', ':', 'return... | Returns the minor loss coefficient for a tapered reducer.
Parameters:
ent_pipe_id: Entrance pipe's inner diameter.
exit_pipe_id: Exit pipe's inner diameter.
fitting_angle: Fitting angle between entrance and exit pipes.
re: Reynold's number.
f: Darcy friction factor. | ['Returns', 'the', 'minor', 'loss', 'coefficient', 'for', 'a', 'tapered', 'reducer', '.'] | train | https://github.com/AguaClara/aguaclara/blob/8dd4e734768b166a7fc2b60388a24df2f93783fc/aguaclara/core/head_loss.py#L175-L195 |
4,598 | SheffieldML/GPy | GPy/models/state_space_main.py | DescreteStateSpace._kalman_prediction_step_SVD | def _kalman_prediction_step_SVD(k, p_m , p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,
p_dm = None, p_dP = None):
"""
Desctrete prediction function
Input:
k:int
Iteration No. Starts at 0. Total number of iterations equal t... | python | def _kalman_prediction_step_SVD(k, p_m , p_P, p_dyn_model_callable, calc_grad_log_likelihood=False,
p_dm = None, p_dP = None):
"""
Desctrete prediction function
Input:
k:int
Iteration No. Starts at 0. Total number of iterations equal t... | ['def', '_kalman_prediction_step_SVD', '(', 'k', ',', 'p_m', ',', 'p_P', ',', 'p_dyn_model_callable', ',', 'calc_grad_log_likelihood', '=', 'False', ',', 'p_dm', '=', 'None', ',', 'p_dP', '=', 'None', ')', ':', '# covariance from the previous step and its SVD decomposition', '# p_prev_cov = v * S * V.T', 'Prev_cov', ',... | Desctrete prediction function
Input:
k:int
Iteration No. Starts at 0. Total number of iterations equal to the
number of measurements.
p_m: matrix of size (state_dim, time_series_no)
Mean value from the previous step. For "multiple time se... | ['Desctrete', 'prediction', 'function'] | train | https://github.com/SheffieldML/GPy/blob/54c32d79d289d622fb18b898aee65a2a431d90cf/GPy/models/state_space_main.py#L1357-L1454 |
4,599 | KelSolaar/Foundations | foundations/trace.py | evaluate_trace_request | def evaluate_trace_request(data, tracer=tracer):
"""
Evaluate given string trace request.
Usage::
Umbra -t "{'umbra.engine' : ('.*', 0), 'umbra.preferences' : (r'.*', 0)}"
Umbra -t "['umbra.engine', 'umbra.preferences']"
Umbra -t "'umbra.engine, umbra.preferences"
:param data:... | python | def evaluate_trace_request(data, tracer=tracer):
"""
Evaluate given string trace request.
Usage::
Umbra -t "{'umbra.engine' : ('.*', 0), 'umbra.preferences' : (r'.*', 0)}"
Umbra -t "['umbra.engine', 'umbra.preferences']"
Umbra -t "'umbra.engine, umbra.preferences"
:param data:... | ['def', 'evaluate_trace_request', '(', 'data', ',', 'tracer', '=', 'tracer', ')', ':', 'data', '=', 'ast', '.', 'literal_eval', '(', 'data', ')', 'if', 'isinstance', '(', 'data', ',', 'str', ')', ':', 'modules', '=', 'dict', '.', 'fromkeys', '(', 'map', '(', 'lambda', 'x', ':', 'x', '.', 'strip', '(', ')', ',', 'data',... | Evaluate given string trace request.
Usage::
Umbra -t "{'umbra.engine' : ('.*', 0), 'umbra.preferences' : (r'.*', 0)}"
Umbra -t "['umbra.engine', 'umbra.preferences']"
Umbra -t "'umbra.engine, umbra.preferences"
:param data: Trace request.
:type data: unicode
:param tracer: Tr... | ['Evaluate', 'given', 'string', 'trace', 'request', '.'] | train | https://github.com/KelSolaar/Foundations/blob/5c141330faf09dad70a12bc321f4c564917d0a91/foundations/trace.py#L825-L857 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.