id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
18,484 | def build_match_conditions(doctype, as_condition=True):
import frappe.desk.reportview
return frappe.desk.reportview.build_match_conditions(doctype, as_condition)
| [
"def",
"build_match_conditions",
"(",
"doctype",
",",
"as_condition",
"=",
"True",
")",
":",
"import",
"frappe",
".",
"desk",
".",
"reportview",
"return",
"frappe",
".",
"desk",
".",
"reportview",
".",
"build_match_conditions",
"(",
"doctype",
",",
"as_condition",
")"
] | return match for given doctype as list or sql . | train | false |
18,485 | def test_VAE_cost():
yaml_src_path = os.path.join(os.path.dirname(__file__), 'test_vae_cost_vae_criterion.yaml')
train_object = yaml_parse.load_path(yaml_src_path)
train_object.main_loop()
| [
"def",
"test_VAE_cost",
"(",
")",
":",
"yaml_src_path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
",",
"'test_vae_cost_vae_criterion.yaml'",
")",
"train_object",
"=",
"yaml_parse",
".",
"load_path",
"(",
"yaml_src_path",
")",
"train_object",
".",
"main_loop",
"(",
")"
] | vae trains properly with the vae cost . | train | false |
18,487 | def _notebook_run(path):
kernel_name = ('python%d' % sys.version_info[0])
this_file_directory = os.path.dirname(__file__)
errors = []
with tempfile.NamedTemporaryFile(suffix='.ipynb', mode='wt') as fout:
with open(path) as f:
nb = nbformat.read(f, as_version=4)
nb.metadata.get('kernelspec', {})['name'] = kernel_name
ep = ExecutePreprocessor(kernel_name=kernel_name, timeout=10)
try:
ep.preprocess(nb, {'metadata': {'path': this_file_directory}})
except CellExecutionError as e:
if ('SKIP' in e.traceback):
print str(e.traceback).split('\n')[(-2)]
else:
raise e
except TimeoutError as e:
print e
finally:
nbformat.write(nb, fout)
return (nb, errors)
| [
"def",
"_notebook_run",
"(",
"path",
")",
":",
"kernel_name",
"=",
"(",
"'python%d'",
"%",
"sys",
".",
"version_info",
"[",
"0",
"]",
")",
"this_file_directory",
"=",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
"errors",
"=",
"[",
"]",
"with",
"tempfile",
".",
"NamedTemporaryFile",
"(",
"suffix",
"=",
"'.ipynb'",
",",
"mode",
"=",
"'wt'",
")",
"as",
"fout",
":",
"with",
"open",
"(",
"path",
")",
"as",
"f",
":",
"nb",
"=",
"nbformat",
".",
"read",
"(",
"f",
",",
"as_version",
"=",
"4",
")",
"nb",
".",
"metadata",
".",
"get",
"(",
"'kernelspec'",
",",
"{",
"}",
")",
"[",
"'name'",
"]",
"=",
"kernel_name",
"ep",
"=",
"ExecutePreprocessor",
"(",
"kernel_name",
"=",
"kernel_name",
",",
"timeout",
"=",
"10",
")",
"try",
":",
"ep",
".",
"preprocess",
"(",
"nb",
",",
"{",
"'metadata'",
":",
"{",
"'path'",
":",
"this_file_directory",
"}",
"}",
")",
"except",
"CellExecutionError",
"as",
"e",
":",
"if",
"(",
"'SKIP'",
"in",
"e",
".",
"traceback",
")",
":",
"print",
"str",
"(",
"e",
".",
"traceback",
")",
".",
"split",
"(",
"'\\n'",
")",
"[",
"(",
"-",
"2",
")",
"]",
"else",
":",
"raise",
"e",
"except",
"TimeoutError",
"as",
"e",
":",
"print",
"e",
"finally",
":",
"nbformat",
".",
"write",
"(",
"nb",
",",
"fout",
")",
"return",
"(",
"nb",
",",
"errors",
")"
] | execute a notebook via nbconvert and collect output . | train | false |
18,489 | @profiler.trace
@memoized
def tenant_quota_usages(request, tenant_id=None):
if (not tenant_id):
tenant_id = request.user.project_id
disabled_quotas = get_disabled_quotas(request)
usages = QuotaUsage()
for quota in get_tenant_quota_data(request, disabled_quotas=disabled_quotas, tenant_id=tenant_id):
usages.add_quota(quota)
_get_tenant_compute_usages(request, usages, disabled_quotas, tenant_id)
_get_tenant_network_usages(request, usages, disabled_quotas, tenant_id)
_get_tenant_volume_usages(request, usages, disabled_quotas, tenant_id)
return usages
| [
"@",
"profiler",
".",
"trace",
"@",
"memoized",
"def",
"tenant_quota_usages",
"(",
"request",
",",
"tenant_id",
"=",
"None",
")",
":",
"if",
"(",
"not",
"tenant_id",
")",
":",
"tenant_id",
"=",
"request",
".",
"user",
".",
"project_id",
"disabled_quotas",
"=",
"get_disabled_quotas",
"(",
"request",
")",
"usages",
"=",
"QuotaUsage",
"(",
")",
"for",
"quota",
"in",
"get_tenant_quota_data",
"(",
"request",
",",
"disabled_quotas",
"=",
"disabled_quotas",
",",
"tenant_id",
"=",
"tenant_id",
")",
":",
"usages",
".",
"add_quota",
"(",
"quota",
")",
"_get_tenant_compute_usages",
"(",
"request",
",",
"usages",
",",
"disabled_quotas",
",",
"tenant_id",
")",
"_get_tenant_network_usages",
"(",
"request",
",",
"usages",
",",
"disabled_quotas",
",",
"tenant_id",
")",
"_get_tenant_volume_usages",
"(",
"request",
",",
"usages",
",",
"disabled_quotas",
",",
"tenant_id",
")",
"return",
"usages"
] | get our quotas and construct our usage object . | train | false |
18,492 | def _FileMatches(path, excl_regexps):
for r in excl_regexps:
if re.match(r, path):
return False
return True
| [
"def",
"_FileMatches",
"(",
"path",
",",
"excl_regexps",
")",
":",
"for",
"r",
"in",
"excl_regexps",
":",
"if",
"re",
".",
"match",
"(",
"r",
",",
"path",
")",
":",
"return",
"False",
"return",
"True"
] | returns true if the specified path matches none of the specified exclude regular expresions . | train | false |
18,493 | def Canonicalize(node, output=None, **kw):
if output:
apply(_implementation, (node, output.write), kw)
else:
s = StringIO.StringIO()
apply(_implementation, (node, s.write), kw)
return s.getvalue()
| [
"def",
"Canonicalize",
"(",
"node",
",",
"output",
"=",
"None",
",",
"**",
"kw",
")",
":",
"if",
"output",
":",
"apply",
"(",
"_implementation",
",",
"(",
"node",
",",
"output",
".",
"write",
")",
",",
"kw",
")",
"else",
":",
"s",
"=",
"StringIO",
".",
"StringIO",
"(",
")",
"apply",
"(",
"_implementation",
",",
"(",
"node",
",",
"s",
".",
"write",
")",
",",
"kw",
")",
"return",
"s",
".",
"getvalue",
"(",
")"
] | canonicalize -> utf-8 canonicalize a dom document/element node and all descendents . | train | true |
18,494 | def CheckTransaction(request_trusted, request_app_id, transaction):
assert isinstance(transaction, datastore_pb.Transaction)
CheckAppId(request_trusted, request_app_id, transaction.app())
| [
"def",
"CheckTransaction",
"(",
"request_trusted",
",",
"request_app_id",
",",
"transaction",
")",
":",
"assert",
"isinstance",
"(",
"transaction",
",",
"datastore_pb",
".",
"Transaction",
")",
"CheckAppId",
"(",
"request_trusted",
",",
"request_app_id",
",",
"transaction",
".",
"app",
"(",
")",
")"
] | check that this transaction is valid . | train | false |
18,495 | def DoesTestHaveLabels(cls, labels):
labels = set(labels)
for name in dir(cls):
if name.startswith('test'):
item = getattr(cls, name, None)
if labels.intersection(getattr(item, 'labels', set(['small']))):
return True
return False
| [
"def",
"DoesTestHaveLabels",
"(",
"cls",
",",
"labels",
")",
":",
"labels",
"=",
"set",
"(",
"labels",
")",
"for",
"name",
"in",
"dir",
"(",
"cls",
")",
":",
"if",
"name",
".",
"startswith",
"(",
"'test'",
")",
":",
"item",
"=",
"getattr",
"(",
"cls",
",",
"name",
",",
"None",
")",
"if",
"labels",
".",
"intersection",
"(",
"getattr",
"(",
"item",
",",
"'labels'",
",",
"set",
"(",
"[",
"'small'",
"]",
")",
")",
")",
":",
"return",
"True",
"return",
"False"
] | returns true if any tests in cls have any of the labels . | train | false |
18,496 | def find_loader(fullname):
for importer in iter_importers(fullname):
loader = importer.find_module(fullname)
if (loader is not None):
return loader
return None
| [
"def",
"find_loader",
"(",
"fullname",
")",
":",
"for",
"importer",
"in",
"iter_importers",
"(",
"fullname",
")",
":",
"loader",
"=",
"importer",
".",
"find_module",
"(",
"fullname",
")",
"if",
"(",
"loader",
"is",
"not",
"None",
")",
":",
"return",
"loader",
"return",
"None"
] | find a pep 302 "loader" object for fullname if fullname contains dots . | train | true |
18,497 | def prepareFailedName(release):
fixed = urllib.unquote(release)
if fixed.endswith('.nzb'):
fixed = fixed.rpartition('.')[0]
fixed = re.sub('[\\.\\-\\+\\ ]', '_', fixed)
fixed = ss(fixed)
return fixed
| [
"def",
"prepareFailedName",
"(",
"release",
")",
":",
"fixed",
"=",
"urllib",
".",
"unquote",
"(",
"release",
")",
"if",
"fixed",
".",
"endswith",
"(",
"'.nzb'",
")",
":",
"fixed",
"=",
"fixed",
".",
"rpartition",
"(",
"'.'",
")",
"[",
"0",
"]",
"fixed",
"=",
"re",
".",
"sub",
"(",
"'[\\\\.\\\\-\\\\+\\\\ ]'",
",",
"'_'",
",",
"fixed",
")",
"fixed",
"=",
"ss",
"(",
"fixed",
")",
"return",
"fixed"
] | standardizes release name for failed db . | train | false |
18,498 | def weight_variable(shape):
initial = tf.zeros(shape)
return tf.Variable(initial)
| [
"def",
"weight_variable",
"(",
"shape",
")",
":",
"initial",
"=",
"tf",
".",
"zeros",
"(",
"shape",
")",
"return",
"tf",
".",
"Variable",
"(",
"initial",
")"
] | helper function to create a weight variable initialized with a normal distribution parameters shape : list size of weight variable . | train | false |
18,499 | def reload_modules():
return True
| [
"def",
"reload_modules",
"(",
")",
":",
"return",
"True"
] | tell the minion to reload the execution modules cli example: . | train | false |
18,501 | def _get_enabled_provider(provider_id):
enabled_provider = provider.Registry.get(provider_id)
if (not enabled_provider):
raise ValueError(('Provider %s not enabled' % provider_id))
return enabled_provider
| [
"def",
"_get_enabled_provider",
"(",
"provider_id",
")",
":",
"enabled_provider",
"=",
"provider",
".",
"Registry",
".",
"get",
"(",
"provider_id",
")",
"if",
"(",
"not",
"enabled_provider",
")",
":",
"raise",
"ValueError",
"(",
"(",
"'Provider %s not enabled'",
"%",
"provider_id",
")",
")",
"return",
"enabled_provider"
] | gets an enabled provider by its provider_id member or throws . | train | false |
18,502 | def compareYAscending(point, pointOther):
if (point.y < pointOther.y):
return (-1)
return int((point.y > pointOther.y))
| [
"def",
"compareYAscending",
"(",
"point",
",",
"pointOther",
")",
":",
"if",
"(",
"point",
".",
"y",
"<",
"pointOther",
".",
"y",
")",
":",
"return",
"(",
"-",
"1",
")",
"return",
"int",
"(",
"(",
"point",
".",
"y",
">",
"pointOther",
".",
"y",
")",
")"
] | get comparison in order to sort points in ascending y . | train | false |
18,503 | def get_params_type(descriptor):
params = descriptor.split(')')[0][1:].split()
if params:
return [param for param in params]
return []
| [
"def",
"get_params_type",
"(",
"descriptor",
")",
":",
"params",
"=",
"descriptor",
".",
"split",
"(",
"')'",
")",
"[",
"0",
"]",
"[",
"1",
":",
"]",
".",
"split",
"(",
")",
"if",
"params",
":",
"return",
"[",
"param",
"for",
"param",
"in",
"params",
"]",
"return",
"[",
"]"
] | return the parameters type of a descriptor (e . | train | true |
18,505 | def escape_identifier(text, reg=KWD_RE):
if (not text):
return '_'
if text[0].isdigit():
text = ('_' + text)
return reg.sub('\\1_', text)
| [
"def",
"escape_identifier",
"(",
"text",
",",
"reg",
"=",
"KWD_RE",
")",
":",
"if",
"(",
"not",
"text",
")",
":",
"return",
"'_'",
"if",
"text",
"[",
"0",
"]",
".",
"isdigit",
"(",
")",
":",
"text",
"=",
"(",
"'_'",
"+",
"text",
")",
"return",
"reg",
".",
"sub",
"(",
"'\\\\1_'",
",",
"text",
")"
] | escape partial c identifiers so they can be used as attributes/arguments . | train | true |
18,506 | def preview_loading(request):
return HttpResponse(u'<html><head><title></title></head><body></body></html>')
| [
"def",
"preview_loading",
"(",
"request",
")",
":",
"return",
"HttpResponse",
"(",
"u'<html><head><title></title></head><body></body></html>'",
")"
] | this page is blank . | train | false |
18,507 | def test_select_birthday_widget():
assert (SelectBirthdayWidget.FORMAT_CHOICES['%d'] == [(x, str(x)) for x in range(1, 32)])
assert (SelectBirthdayWidget.FORMAT_CHOICES['%m'] == [(x, str(x)) for x in range(1, 13)])
assert (SelectBirthdayWidget.FORMAT_CLASSES == {'%d': 'select_date_day', '%m': 'select_date_month', '%Y': 'select_date_year'})
select_birthday_widget = SelectBirthdayWidget(years=[0, 1])
assert (select_birthday_widget.FORMAT_CHOICES['%Y'] == [(0, '0'), (1, '1')])
class Field(object, ):
id = 'world'
name = 'helloWorld'
format = '%d %m %Y'
data = None
html = select_birthday_widget(field=Field(), surrounded_div='test-div')
assert ('world' in html)
assert ('helloWorld' in html)
assert ('class="select_date_day"' in html)
assert ('class="select_date_month"' in html)
assert ('class="select_date_year"' in html)
assert ('<div class="test-div">' in html)
| [
"def",
"test_select_birthday_widget",
"(",
")",
":",
"assert",
"(",
"SelectBirthdayWidget",
".",
"FORMAT_CHOICES",
"[",
"'%d'",
"]",
"==",
"[",
"(",
"x",
",",
"str",
"(",
"x",
")",
")",
"for",
"x",
"in",
"range",
"(",
"1",
",",
"32",
")",
"]",
")",
"assert",
"(",
"SelectBirthdayWidget",
".",
"FORMAT_CHOICES",
"[",
"'%m'",
"]",
"==",
"[",
"(",
"x",
",",
"str",
"(",
"x",
")",
")",
"for",
"x",
"in",
"range",
"(",
"1",
",",
"13",
")",
"]",
")",
"assert",
"(",
"SelectBirthdayWidget",
".",
"FORMAT_CLASSES",
"==",
"{",
"'%d'",
":",
"'select_date_day'",
",",
"'%m'",
":",
"'select_date_month'",
",",
"'%Y'",
":",
"'select_date_year'",
"}",
")",
"select_birthday_widget",
"=",
"SelectBirthdayWidget",
"(",
"years",
"=",
"[",
"0",
",",
"1",
"]",
")",
"assert",
"(",
"select_birthday_widget",
".",
"FORMAT_CHOICES",
"[",
"'%Y'",
"]",
"==",
"[",
"(",
"0",
",",
"'0'",
")",
",",
"(",
"1",
",",
"'1'",
")",
"]",
")",
"class",
"Field",
"(",
"object",
",",
")",
":",
"id",
"=",
"'world'",
"name",
"=",
"'helloWorld'",
"format",
"=",
"'%d %m %Y'",
"data",
"=",
"None",
"html",
"=",
"select_birthday_widget",
"(",
"field",
"=",
"Field",
"(",
")",
",",
"surrounded_div",
"=",
"'test-div'",
")",
"assert",
"(",
"'world'",
"in",
"html",
")",
"assert",
"(",
"'helloWorld'",
"in",
"html",
")",
"assert",
"(",
"'class=\"select_date_day\"'",
"in",
"html",
")",
"assert",
"(",
"'class=\"select_date_month\"'",
"in",
"html",
")",
"assert",
"(",
"'class=\"select_date_year\"'",
"in",
"html",
")",
"assert",
"(",
"'<div class=\"test-div\">'",
"in",
"html",
")"
] | test the selectdatewidget . | train | false |
18,509 | def fontifyPython(document):
def matcher(node):
return ((node.nodeName == 'pre') and node.hasAttribute('class') and (node.getAttribute('class') == 'python'))
for node in domhelpers.findElements(document, matcher):
fontifyPythonNode(node)
| [
"def",
"fontifyPython",
"(",
"document",
")",
":",
"def",
"matcher",
"(",
"node",
")",
":",
"return",
"(",
"(",
"node",
".",
"nodeName",
"==",
"'pre'",
")",
"and",
"node",
".",
"hasAttribute",
"(",
"'class'",
")",
"and",
"(",
"node",
".",
"getAttribute",
"(",
"'class'",
")",
"==",
"'python'",
")",
")",
"for",
"node",
"in",
"domhelpers",
".",
"findElements",
"(",
"document",
",",
"matcher",
")",
":",
"fontifyPythonNode",
"(",
"node",
")"
] | syntax color any node in the given document which contains a python source listing . | train | false |
18,510 | def solrctl():
for dirname in os.environ.get('PATH', '').split(os.path.pathsep):
path = os.path.join(dirname, 'solrctl')
if os.path.exists(path):
return path
return None
| [
"def",
"solrctl",
"(",
")",
":",
"for",
"dirname",
"in",
"os",
".",
"environ",
".",
"get",
"(",
"'PATH'",
",",
"''",
")",
".",
"split",
"(",
"os",
".",
"path",
".",
"pathsep",
")",
":",
"path",
"=",
"os",
".",
"path",
".",
"join",
"(",
"dirname",
",",
"'solrctl'",
")",
"if",
"os",
".",
"path",
".",
"exists",
"(",
"path",
")",
":",
"return",
"path",
"return",
"None"
] | solrctl path . | train | false |
18,511 | def attach_const_node(node, name, value):
if (not (name in node.special_attributes)):
_attach_local_node(node, const_factory(value), name)
| [
"def",
"attach_const_node",
"(",
"node",
",",
"name",
",",
"value",
")",
":",
"if",
"(",
"not",
"(",
"name",
"in",
"node",
".",
"special_attributes",
")",
")",
":",
"_attach_local_node",
"(",
"node",
",",
"const_factory",
"(",
"value",
")",
",",
"name",
")"
] | create a const node and register it in the locals of the given node with the specified name . | train | true |
18,513 | def image_get_all(context, filters=None, marker=None, limit=None, sort_key=None, sort_dir=None, member_status='accepted', is_public=None, admin_as_user=False, return_tag=False, v1_mode=False):
sort_key = (['created_at'] if (not sort_key) else sort_key)
default_sort_dir = 'desc'
if (not sort_dir):
sort_dir = ([default_sort_dir] * len(sort_key))
elif (len(sort_dir) == 1):
default_sort_dir = sort_dir[0]
sort_dir *= len(sort_key)
filters = (filters or {})
visibility = filters.pop('visibility', None)
showing_deleted = (('changes-since' in filters) or filters.get('deleted', False))
(img_cond, prop_cond, tag_cond) = _make_conditions_from_filters(filters, is_public)
query = _select_images_query(context, img_cond, admin_as_user, member_status, visibility)
if (visibility is not None):
query = query.filter((models.Image.visibility == visibility))
elif (context.owner is None):
query = query.filter((models.Image.visibility != 'community'))
else:
community_filters = [(models.Image.owner == context.owner), (models.Image.visibility != 'community')]
query = query.filter(sa_sql.or_(*community_filters))
if prop_cond:
for prop_condition in prop_cond:
query = query.join(models.ImageProperty, aliased=True).filter(sa_sql.and_(*prop_condition))
if tag_cond:
for tag_condition in tag_cond:
query = query.join(models.ImageTag, aliased=True).filter(sa_sql.and_(*tag_condition))
marker_image = None
if (marker is not None):
marker_image = _image_get(context, marker, force_show_deleted=showing_deleted)
for key in ['created_at', 'id']:
if (key not in sort_key):
sort_key.append(key)
sort_dir.append(default_sort_dir)
query = _paginate_query(query, models.Image, limit, sort_key, marker=marker_image, sort_dir=None, sort_dirs=sort_dir)
query = query.options(sa_orm.joinedload(models.Image.properties)).options(sa_orm.joinedload(models.Image.locations))
if return_tag:
query = query.options(sa_orm.joinedload(models.Image.tags))
images = []
for image in query.all():
image_dict = image.to_dict()
image_dict = _normalize_locations(context, image_dict, force_show_deleted=showing_deleted)
if return_tag:
image_dict = _normalize_tags(image_dict)
if v1_mode:
image_dict = db_utils.mutate_image_dict_to_v1(image_dict)
images.append(image_dict)
return images
| [
"def",
"image_get_all",
"(",
"context",
",",
"filters",
"=",
"None",
",",
"marker",
"=",
"None",
",",
"limit",
"=",
"None",
",",
"sort_key",
"=",
"None",
",",
"sort_dir",
"=",
"None",
",",
"member_status",
"=",
"'accepted'",
",",
"is_public",
"=",
"None",
",",
"admin_as_user",
"=",
"False",
",",
"return_tag",
"=",
"False",
",",
"v1_mode",
"=",
"False",
")",
":",
"sort_key",
"=",
"(",
"[",
"'created_at'",
"]",
"if",
"(",
"not",
"sort_key",
")",
"else",
"sort_key",
")",
"default_sort_dir",
"=",
"'desc'",
"if",
"(",
"not",
"sort_dir",
")",
":",
"sort_dir",
"=",
"(",
"[",
"default_sort_dir",
"]",
"*",
"len",
"(",
"sort_key",
")",
")",
"elif",
"(",
"len",
"(",
"sort_dir",
")",
"==",
"1",
")",
":",
"default_sort_dir",
"=",
"sort_dir",
"[",
"0",
"]",
"sort_dir",
"*=",
"len",
"(",
"sort_key",
")",
"filters",
"=",
"(",
"filters",
"or",
"{",
"}",
")",
"visibility",
"=",
"filters",
".",
"pop",
"(",
"'visibility'",
",",
"None",
")",
"showing_deleted",
"=",
"(",
"(",
"'changes-since'",
"in",
"filters",
")",
"or",
"filters",
".",
"get",
"(",
"'deleted'",
",",
"False",
")",
")",
"(",
"img_cond",
",",
"prop_cond",
",",
"tag_cond",
")",
"=",
"_make_conditions_from_filters",
"(",
"filters",
",",
"is_public",
")",
"query",
"=",
"_select_images_query",
"(",
"context",
",",
"img_cond",
",",
"admin_as_user",
",",
"member_status",
",",
"visibility",
")",
"if",
"(",
"visibility",
"is",
"not",
"None",
")",
":",
"query",
"=",
"query",
".",
"filter",
"(",
"(",
"models",
".",
"Image",
".",
"visibility",
"==",
"visibility",
")",
")",
"elif",
"(",
"context",
".",
"owner",
"is",
"None",
")",
":",
"query",
"=",
"query",
".",
"filter",
"(",
"(",
"models",
".",
"Image",
".",
"visibility",
"!=",
"'community'",
")",
")",
"else",
":",
"community_filters",
"=",
"[",
"(",
"models",
".",
"Image",
".",
"owner",
"==",
"context",
".",
"owner",
")",
",",
"(",
"models",
".",
"Image",
".",
"visibility",
"!=",
"'community'",
")",
"]",
"query",
"=",
"query",
".",
"filter",
"(",
"sa_sql",
".",
"or_",
"(",
"*",
"community_filters",
")",
")",
"if",
"prop_cond",
":",
"for",
"prop_condition",
"in",
"prop_cond",
":",
"query",
"=",
"query",
".",
"join",
"(",
"models",
".",
"ImageProperty",
",",
"aliased",
"=",
"True",
")",
".",
"filter",
"(",
"sa_sql",
".",
"and_",
"(",
"*",
"prop_condition",
")",
")",
"if",
"tag_cond",
":",
"for",
"tag_condition",
"in",
"tag_cond",
":",
"query",
"=",
"query",
".",
"join",
"(",
"models",
".",
"ImageTag",
",",
"aliased",
"=",
"True",
")",
".",
"filter",
"(",
"sa_sql",
".",
"and_",
"(",
"*",
"tag_condition",
")",
")",
"marker_image",
"=",
"None",
"if",
"(",
"marker",
"is",
"not",
"None",
")",
":",
"marker_image",
"=",
"_image_get",
"(",
"context",
",",
"marker",
",",
"force_show_deleted",
"=",
"showing_deleted",
")",
"for",
"key",
"in",
"[",
"'created_at'",
",",
"'id'",
"]",
":",
"if",
"(",
"key",
"not",
"in",
"sort_key",
")",
":",
"sort_key",
".",
"append",
"(",
"key",
")",
"sort_dir",
".",
"append",
"(",
"default_sort_dir",
")",
"query",
"=",
"_paginate_query",
"(",
"query",
",",
"models",
".",
"Image",
",",
"limit",
",",
"sort_key",
",",
"marker",
"=",
"marker_image",
",",
"sort_dir",
"=",
"None",
",",
"sort_dirs",
"=",
"sort_dir",
")",
"query",
"=",
"query",
".",
"options",
"(",
"sa_orm",
".",
"joinedload",
"(",
"models",
".",
"Image",
".",
"properties",
")",
")",
".",
"options",
"(",
"sa_orm",
".",
"joinedload",
"(",
"models",
".",
"Image",
".",
"locations",
")",
")",
"if",
"return_tag",
":",
"query",
"=",
"query",
".",
"options",
"(",
"sa_orm",
".",
"joinedload",
"(",
"models",
".",
"Image",
".",
"tags",
")",
")",
"images",
"=",
"[",
"]",
"for",
"image",
"in",
"query",
".",
"all",
"(",
")",
":",
"image_dict",
"=",
"image",
".",
"to_dict",
"(",
")",
"image_dict",
"=",
"_normalize_locations",
"(",
"context",
",",
"image_dict",
",",
"force_show_deleted",
"=",
"showing_deleted",
")",
"if",
"return_tag",
":",
"image_dict",
"=",
"_normalize_tags",
"(",
"image_dict",
")",
"if",
"v1_mode",
":",
"image_dict",
"=",
"db_utils",
".",
"mutate_image_dict_to_v1",
"(",
"image_dict",
")",
"images",
".",
"append",
"(",
"image_dict",
")",
"return",
"images"
] | get all images that match zero or more filters . | train | false |
18,515 | def get_profiler_results_dir(autodir):
return os.path.join(autodir, 'results', 'default', 'profiler_sync', 'profiling')
| [
"def",
"get_profiler_results_dir",
"(",
"autodir",
")",
":",
"return",
"os",
".",
"path",
".",
"join",
"(",
"autodir",
",",
"'results'",
",",
"'default'",
",",
"'profiler_sync'",
",",
"'profiling'",
")"
] | given the directory of the autotest_remote client used to run a profiler . | train | false |
18,516 | def fitsopen(name, mode='readonly', memmap=None, save_backup=False, cache=True, lazy_load_hdus=None, **kwargs):
from .. import conf
if (memmap is None):
memmap = (None if conf.use_memmap else False)
else:
memmap = bool(memmap)
if (lazy_load_hdus is None):
lazy_load_hdus = conf.lazy_load_hdus
else:
lazy_load_hdus = bool(lazy_load_hdus)
if (('uint16' in kwargs) and ('uint' not in kwargs)):
kwargs['uint'] = kwargs['uint16']
del kwargs['uint16']
warnings.warn('The uint16 keyword argument is deprecated since v1.1.0. Use the uint argument instead.', AstropyDeprecationWarning)
if ('uint' not in kwargs):
kwargs['uint'] = conf.enable_uint
if (not name):
raise ValueError('Empty filename: {}'.format(repr(name)))
return HDUList.fromfile(name, mode, memmap, save_backup, cache, lazy_load_hdus, **kwargs)
| [
"def",
"fitsopen",
"(",
"name",
",",
"mode",
"=",
"'readonly'",
",",
"memmap",
"=",
"None",
",",
"save_backup",
"=",
"False",
",",
"cache",
"=",
"True",
",",
"lazy_load_hdus",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"from",
".",
".",
"import",
"conf",
"if",
"(",
"memmap",
"is",
"None",
")",
":",
"memmap",
"=",
"(",
"None",
"if",
"conf",
".",
"use_memmap",
"else",
"False",
")",
"else",
":",
"memmap",
"=",
"bool",
"(",
"memmap",
")",
"if",
"(",
"lazy_load_hdus",
"is",
"None",
")",
":",
"lazy_load_hdus",
"=",
"conf",
".",
"lazy_load_hdus",
"else",
":",
"lazy_load_hdus",
"=",
"bool",
"(",
"lazy_load_hdus",
")",
"if",
"(",
"(",
"'uint16'",
"in",
"kwargs",
")",
"and",
"(",
"'uint'",
"not",
"in",
"kwargs",
")",
")",
":",
"kwargs",
"[",
"'uint'",
"]",
"=",
"kwargs",
"[",
"'uint16'",
"]",
"del",
"kwargs",
"[",
"'uint16'",
"]",
"warnings",
".",
"warn",
"(",
"'The uint16 keyword argument is deprecated since v1.1.0. Use the uint argument instead.'",
",",
"AstropyDeprecationWarning",
")",
"if",
"(",
"'uint'",
"not",
"in",
"kwargs",
")",
":",
"kwargs",
"[",
"'uint'",
"]",
"=",
"conf",
".",
"enable_uint",
"if",
"(",
"not",
"name",
")",
":",
"raise",
"ValueError",
"(",
"'Empty filename: {}'",
".",
"format",
"(",
"repr",
"(",
"name",
")",
")",
")",
"return",
"HDUList",
".",
"fromfile",
"(",
"name",
",",
"mode",
",",
"memmap",
",",
"save_backup",
",",
"cache",
",",
"lazy_load_hdus",
",",
"**",
"kwargs",
")"
] | factory function to open a fits file and return an hdulist object . | train | false |
18,517 | def seconds_to_time(x):
t = int((x * (10 ** 6)))
ms = (t % (10 ** 6))
t = (t // (10 ** 6))
s = (t % 60)
t = (t // 60)
m = (t % 60)
t = (t // 60)
h = t
return time(h, m, s, ms)
| [
"def",
"seconds_to_time",
"(",
"x",
")",
":",
"t",
"=",
"int",
"(",
"(",
"x",
"*",
"(",
"10",
"**",
"6",
")",
")",
")",
"ms",
"=",
"(",
"t",
"%",
"(",
"10",
"**",
"6",
")",
")",
"t",
"=",
"(",
"t",
"//",
"(",
"10",
"**",
"6",
")",
")",
"s",
"=",
"(",
"t",
"%",
"60",
")",
"t",
"=",
"(",
"t",
"//",
"60",
")",
"m",
"=",
"(",
"t",
"%",
"60",
")",
"t",
"=",
"(",
"t",
"//",
"60",
")",
"h",
"=",
"t",
"return",
"time",
"(",
"h",
",",
"m",
",",
"s",
",",
"ms",
")"
] | convert a number of second into a time . | train | true |
18,518 | def task_update(context, task_id, values, session=None):
session = (session or get_session())
with session.begin():
task_info_values = _pop_task_info_values(values)
task_ref = _task_get(context, task_id, session)
_drop_protected_attrs(models.Task, values)
values['updated_at'] = timeutils.utcnow()
_task_update(context, task_ref, values, session)
if task_info_values:
_task_info_update(context, task_id, task_info_values, session)
return task_get(context, task_id, session)
| [
"def",
"task_update",
"(",
"context",
",",
"task_id",
",",
"values",
",",
"session",
"=",
"None",
")",
":",
"session",
"=",
"(",
"session",
"or",
"get_session",
"(",
")",
")",
"with",
"session",
".",
"begin",
"(",
")",
":",
"task_info_values",
"=",
"_pop_task_info_values",
"(",
"values",
")",
"task_ref",
"=",
"_task_get",
"(",
"context",
",",
"task_id",
",",
"session",
")",
"_drop_protected_attrs",
"(",
"models",
".",
"Task",
",",
"values",
")",
"values",
"[",
"'updated_at'",
"]",
"=",
"timeutils",
".",
"utcnow",
"(",
")",
"_task_update",
"(",
"context",
",",
"task_ref",
",",
"values",
",",
"session",
")",
"if",
"task_info_values",
":",
"_task_info_update",
"(",
"context",
",",
"task_id",
",",
"task_info_values",
",",
"session",
")",
"return",
"task_get",
"(",
"context",
",",
"task_id",
",",
"session",
")"
] | update a task object . | train | false |
18,519 | def convert_ListProperty(model, prop, kwargs):
return None
| [
"def",
"convert_ListProperty",
"(",
"model",
",",
"prop",
",",
"kwargs",
")",
":",
"return",
"None"
] | returns a form field for a db . | train | false |
18,520 | def ssh_command_as_root(ssh_connection, cmd, check_exit_code=True):
LOG.debug((_('Running cmd (SSH-as-root): %s') % cmd))
chan = ssh_connection._transport.open_session()
chan.exec_command('ioscli oem_setup_env')
bufsize = (-1)
stdin = chan.makefile('wb', bufsize)
stdout = chan.makefile('rb', bufsize)
stderr = chan.makefile_stderr('rb', bufsize)
stdin.write(('%s\n%s\n' % (cmd, 'exit')))
stdin.flush()
exit_status = chan.recv_exit_status()
if (exit_status != (-1)):
LOG.debug((_('Result was %s') % exit_status))
if (check_exit_code and (exit_status != 0)):
raise nova_exception.ProcessExecutionError(exit_code=exit_status, stdout=stdout, stderr=stderr, cmd=''.join(cmd))
return (stdout, stderr)
| [
"def",
"ssh_command_as_root",
"(",
"ssh_connection",
",",
"cmd",
",",
"check_exit_code",
"=",
"True",
")",
":",
"LOG",
".",
"debug",
"(",
"(",
"_",
"(",
"'Running cmd (SSH-as-root): %s'",
")",
"%",
"cmd",
")",
")",
"chan",
"=",
"ssh_connection",
".",
"_transport",
".",
"open_session",
"(",
")",
"chan",
".",
"exec_command",
"(",
"'ioscli oem_setup_env'",
")",
"bufsize",
"=",
"(",
"-",
"1",
")",
"stdin",
"=",
"chan",
".",
"makefile",
"(",
"'wb'",
",",
"bufsize",
")",
"stdout",
"=",
"chan",
".",
"makefile",
"(",
"'rb'",
",",
"bufsize",
")",
"stderr",
"=",
"chan",
".",
"makefile_stderr",
"(",
"'rb'",
",",
"bufsize",
")",
"stdin",
".",
"write",
"(",
"(",
"'%s\\n%s\\n'",
"%",
"(",
"cmd",
",",
"'exit'",
")",
")",
")",
"stdin",
".",
"flush",
"(",
")",
"exit_status",
"=",
"chan",
".",
"recv_exit_status",
"(",
")",
"if",
"(",
"exit_status",
"!=",
"(",
"-",
"1",
")",
")",
":",
"LOG",
".",
"debug",
"(",
"(",
"_",
"(",
"'Result was %s'",
")",
"%",
"exit_status",
")",
")",
"if",
"(",
"check_exit_code",
"and",
"(",
"exit_status",
"!=",
"0",
")",
")",
":",
"raise",
"nova_exception",
".",
"ProcessExecutionError",
"(",
"exit_code",
"=",
"exit_status",
",",
"stdout",
"=",
"stdout",
",",
"stderr",
"=",
"stderr",
",",
"cmd",
"=",
"''",
".",
"join",
"(",
"cmd",
")",
")",
"return",
"(",
"stdout",
",",
"stderr",
")"
] | method to execute remote command as root . | train | false |
18,521 | def VaOutput(format, args):
text = (format % args)
if (_Level > 0):
indent = (' DCTB ' * _Level)
lines = text.split('\n')
for i in range(len(lines)):
if (lines[i] and (lines[i][0] != '#')):
lines[i] = (indent + lines[i])
text = '\n'.join(lines)
_File.write((text + '\n'))
| [
"def",
"VaOutput",
"(",
"format",
",",
"args",
")",
":",
"text",
"=",
"(",
"format",
"%",
"args",
")",
"if",
"(",
"_Level",
">",
"0",
")",
":",
"indent",
"=",
"(",
"' DCTB '",
"*",
"_Level",
")",
"lines",
"=",
"text",
".",
"split",
"(",
"'\\n'",
")",
"for",
"i",
"in",
"range",
"(",
"len",
"(",
"lines",
")",
")",
":",
"if",
"(",
"lines",
"[",
"i",
"]",
"and",
"(",
"lines",
"[",
"i",
"]",
"[",
"0",
"]",
"!=",
"'#'",
")",
")",
":",
"lines",
"[",
"i",
"]",
"=",
"(",
"indent",
"+",
"lines",
"[",
"i",
"]",
")",
"text",
"=",
"'\\n'",
".",
"join",
"(",
"lines",
")",
"_File",
".",
"write",
"(",
"(",
"text",
"+",
"'\\n'",
")",
")"
] | call this with a format string and argument tuple for the format . | train | false |
18,523 | def build_pillar_data(options):
pillar = {'test_transport': options.test_transport, 'cloud_only': options.cloud_only, 'with_coverage': (options.test_without_coverage is False)}
if (options.test_git_commit is not None):
pillar['test_git_commit'] = options.test_git_commit
if (options.test_git_url is not None):
pillar['test_git_url'] = options.test_git_url
if (options.bootstrap_salt_url is not None):
pillar['bootstrap_salt_url'] = options.bootstrap_salt_url
if (options.bootstrap_salt_commit is not None):
pillar['bootstrap_salt_commit'] = options.bootstrap_salt_commit
if options.package_source_dir:
pillar['package_source_dir'] = options.package_source_dir
if options.package_build_dir:
pillar['package_build_dir'] = options.package_build_dir
if options.package_artifact_dir:
pillar['package_artifact_dir'] = options.package_artifact_dir
if options.pillar:
pillar.update(dict(options.pillar))
return yaml.dump(pillar, default_flow_style=True, indent=0, width=sys.maxint).rstrip()
| [
"def",
"build_pillar_data",
"(",
"options",
")",
":",
"pillar",
"=",
"{",
"'test_transport'",
":",
"options",
".",
"test_transport",
",",
"'cloud_only'",
":",
"options",
".",
"cloud_only",
",",
"'with_coverage'",
":",
"(",
"options",
".",
"test_without_coverage",
"is",
"False",
")",
"}",
"if",
"(",
"options",
".",
"test_git_commit",
"is",
"not",
"None",
")",
":",
"pillar",
"[",
"'test_git_commit'",
"]",
"=",
"options",
".",
"test_git_commit",
"if",
"(",
"options",
".",
"test_git_url",
"is",
"not",
"None",
")",
":",
"pillar",
"[",
"'test_git_url'",
"]",
"=",
"options",
".",
"test_git_url",
"if",
"(",
"options",
".",
"bootstrap_salt_url",
"is",
"not",
"None",
")",
":",
"pillar",
"[",
"'bootstrap_salt_url'",
"]",
"=",
"options",
".",
"bootstrap_salt_url",
"if",
"(",
"options",
".",
"bootstrap_salt_commit",
"is",
"not",
"None",
")",
":",
"pillar",
"[",
"'bootstrap_salt_commit'",
"]",
"=",
"options",
".",
"bootstrap_salt_commit",
"if",
"options",
".",
"package_source_dir",
":",
"pillar",
"[",
"'package_source_dir'",
"]",
"=",
"options",
".",
"package_source_dir",
"if",
"options",
".",
"package_build_dir",
":",
"pillar",
"[",
"'package_build_dir'",
"]",
"=",
"options",
".",
"package_build_dir",
"if",
"options",
".",
"package_artifact_dir",
":",
"pillar",
"[",
"'package_artifact_dir'",
"]",
"=",
"options",
".",
"package_artifact_dir",
"if",
"options",
".",
"pillar",
":",
"pillar",
".",
"update",
"(",
"dict",
"(",
"options",
".",
"pillar",
")",
")",
"return",
"yaml",
".",
"dump",
"(",
"pillar",
",",
"default_flow_style",
"=",
"True",
",",
"indent",
"=",
"0",
",",
"width",
"=",
"sys",
".",
"maxint",
")",
".",
"rstrip",
"(",
")"
] | build a yaml formatted string to properly pass pillar data . | train | false |
18,524 | def process_tex(lines):
new_lines = []
for line in lines:
line = re.sub('^\\s*\\\\strong{See Also:}\\s*$', '\\paragraph{See Also}', line)
if (line.startswith('\\section{scipy.') or line.startswith('\\subsection{scipy.') or line.startswith('\\subsubsection{scipy.') or line.startswith('\\paragraph{scipy.') or line.startswith('\\subparagraph{scipy.')):
pass
else:
new_lines.append(line)
return new_lines
| [
"def",
"process_tex",
"(",
"lines",
")",
":",
"new_lines",
"=",
"[",
"]",
"for",
"line",
"in",
"lines",
":",
"line",
"=",
"re",
".",
"sub",
"(",
"'^\\\\s*\\\\\\\\strong{See Also:}\\\\s*$'",
",",
"'\\\\paragraph{See Also}'",
",",
"line",
")",
"if",
"(",
"line",
".",
"startswith",
"(",
"'\\\\section{scipy.'",
")",
"or",
"line",
".",
"startswith",
"(",
"'\\\\subsection{scipy.'",
")",
"or",
"line",
".",
"startswith",
"(",
"'\\\\subsubsection{scipy.'",
")",
"or",
"line",
".",
"startswith",
"(",
"'\\\\paragraph{scipy.'",
")",
"or",
"line",
".",
"startswith",
"(",
"'\\\\subparagraph{scipy.'",
")",
")",
":",
"pass",
"else",
":",
"new_lines",
".",
"append",
"(",
"line",
")",
"return",
"new_lines"
] | remove unnecessary section titles from the latex file . | train | false |
18,525 | def sample_func(v):
return (v + v)
| [
"def",
"sample_func",
"(",
"v",
")",
":",
"return",
"(",
"v",
"+",
"v",
")"
] | blah blah . | train | false |
18,526 | def CMakeStringEscape(a):
return a.replace('\\', '\\\\').replace(';', '\\;').replace('"', '\\"')
| [
"def",
"CMakeStringEscape",
"(",
"a",
")",
":",
"return",
"a",
".",
"replace",
"(",
"'\\\\'",
",",
"'\\\\\\\\'",
")",
".",
"replace",
"(",
"';'",
",",
"'\\\\;'",
")",
".",
"replace",
"(",
"'\"'",
",",
"'\\\\\"'",
")"
] | escapes the string a for use inside a cmake string . | train | false |
18,527 | def function(inputs, outputs, updates=[], **kwargs):
if (len(kwargs) > 0):
msg = [('Expected no kwargs, you passed %s' % len(kwargs)), 'kwargs passed to function are ignored with Tensorflow backend']
warnings.warn('\n'.join(msg))
return Function(inputs, outputs, updates=updates)
| [
"def",
"function",
"(",
"inputs",
",",
"outputs",
",",
"updates",
"=",
"[",
"]",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"len",
"(",
"kwargs",
")",
">",
"0",
")",
":",
"msg",
"=",
"[",
"(",
"'Expected no kwargs, you passed %s'",
"%",
"len",
"(",
"kwargs",
")",
")",
",",
"'kwargs passed to function are ignored with Tensorflow backend'",
"]",
"warnings",
".",
"warn",
"(",
"'\\n'",
".",
"join",
"(",
"msg",
")",
")",
"return",
"Function",
"(",
"inputs",
",",
"outputs",
",",
"updates",
"=",
"updates",
")"
] | a wrapper around theano . | train | false |
18,528 | def apply_single_tag_set(tag_set, selection):
def tags_match(server_tags):
for (key, value) in tag_set.items():
if ((key not in server_tags) or (server_tags[key] != value)):
return False
return True
return selection.with_server_descriptions([s for s in selection.server_descriptions if tags_match(s.tags)])
| [
"def",
"apply_single_tag_set",
"(",
"tag_set",
",",
"selection",
")",
":",
"def",
"tags_match",
"(",
"server_tags",
")",
":",
"for",
"(",
"key",
",",
"value",
")",
"in",
"tag_set",
".",
"items",
"(",
")",
":",
"if",
"(",
"(",
"key",
"not",
"in",
"server_tags",
")",
"or",
"(",
"server_tags",
"[",
"key",
"]",
"!=",
"value",
")",
")",
":",
"return",
"False",
"return",
"True",
"return",
"selection",
".",
"with_server_descriptions",
"(",
"[",
"s",
"for",
"s",
"in",
"selection",
".",
"server_descriptions",
"if",
"tags_match",
"(",
"s",
".",
"tags",
")",
"]",
")"
] | all servers matching one tag set . | train | true |
18,529 | @utils.arg('snapshot', metavar='<snapshot>', help='ID of the snapshot.')
@utils.service_type('monitor')
def do_snapshot_show(cs, args):
snapshot = _find_monitor_snapshot(cs, args.snapshot)
_print_monitor_snapshot(snapshot)
| [
"@",
"utils",
".",
"arg",
"(",
"'snapshot'",
",",
"metavar",
"=",
"'<snapshot>'",
",",
"help",
"=",
"'ID of the snapshot.'",
")",
"@",
"utils",
".",
"service_type",
"(",
"'monitor'",
")",
"def",
"do_snapshot_show",
"(",
"cs",
",",
"args",
")",
":",
"snapshot",
"=",
"_find_monitor_snapshot",
"(",
"cs",
",",
"args",
".",
"snapshot",
")",
"_print_monitor_snapshot",
"(",
"snapshot",
")"
] | show details about a snapshot . | train | false |
18,530 | def istest(func):
func.__test__ = True
return func
| [
"def",
"istest",
"(",
"func",
")",
":",
"func",
".",
"__test__",
"=",
"True",
"return",
"func"
] | decorator to mark a function or method as a test . | train | false |
18,531 | def calc_text_angle(start, end):
text_angle = ((start + end) / 2.0)
shift_angles = ((text_angle > (np.pi / 2)) & (text_angle < ((3 * np.pi) / 2)))
text_angle[shift_angles] = (text_angle[shift_angles] + np.pi)
return text_angle
| [
"def",
"calc_text_angle",
"(",
"start",
",",
"end",
")",
":",
"text_angle",
"=",
"(",
"(",
"start",
"+",
"end",
")",
"/",
"2.0",
")",
"shift_angles",
"=",
"(",
"(",
"text_angle",
">",
"(",
"np",
".",
"pi",
"/",
"2",
")",
")",
"&",
"(",
"text_angle",
"<",
"(",
"(",
"3",
"*",
"np",
".",
"pi",
")",
"/",
"2",
")",
")",
")",
"text_angle",
"[",
"shift_angles",
"]",
"=",
"(",
"text_angle",
"[",
"shift_angles",
"]",
"+",
"np",
".",
"pi",
")",
"return",
"text_angle"
] | produce a column of text angle values based on the bounds of the wedge . | train | false |
18,532 | def printc(text, color):
if sys.stdout.isatty():
print (((('\x1b[' + codeCodes[color]) + 'm') + text) + '\x1b[0m')
else:
print text
| [
"def",
"printc",
"(",
"text",
",",
"color",
")",
":",
"if",
"sys",
".",
"stdout",
".",
"isatty",
"(",
")",
":",
"print",
"(",
"(",
"(",
"(",
"'\\x1b['",
"+",
"codeCodes",
"[",
"color",
"]",
")",
"+",
"'m'",
")",
"+",
"text",
")",
"+",
"'\\x1b[0m'",
")",
"else",
":",
"print",
"text"
] | print in color . | train | false |
18,533 | def get_labels(fields, doctype):
labels = []
for key in fields:
key = key.split(u' as ')[0]
if (u'.' in key):
(parenttype, fieldname) = (key.split(u'.')[0][4:(-1)], key.split(u'.')[1].strip(u'`'))
else:
parenttype = doctype
fieldname = fieldname.strip(u'`')
df = frappe.get_meta(parenttype).get_field(fieldname)
label = (df.label if df else fieldname.title())
if (label in labels):
label = ((doctype + u': ') + label)
labels.append(label)
return labels
| [
"def",
"get_labels",
"(",
"fields",
",",
"doctype",
")",
":",
"labels",
"=",
"[",
"]",
"for",
"key",
"in",
"fields",
":",
"key",
"=",
"key",
".",
"split",
"(",
"u' as '",
")",
"[",
"0",
"]",
"if",
"(",
"u'.'",
"in",
"key",
")",
":",
"(",
"parenttype",
",",
"fieldname",
")",
"=",
"(",
"key",
".",
"split",
"(",
"u'.'",
")",
"[",
"0",
"]",
"[",
"4",
":",
"(",
"-",
"1",
")",
"]",
",",
"key",
".",
"split",
"(",
"u'.'",
")",
"[",
"1",
"]",
".",
"strip",
"(",
"u'`'",
")",
")",
"else",
":",
"parenttype",
"=",
"doctype",
"fieldname",
"=",
"fieldname",
".",
"strip",
"(",
"u'`'",
")",
"df",
"=",
"frappe",
".",
"get_meta",
"(",
"parenttype",
")",
".",
"get_field",
"(",
"fieldname",
")",
"label",
"=",
"(",
"df",
".",
"label",
"if",
"df",
"else",
"fieldname",
".",
"title",
"(",
")",
")",
"if",
"(",
"label",
"in",
"labels",
")",
":",
"label",
"=",
"(",
"(",
"doctype",
"+",
"u': '",
")",
"+",
"label",
")",
"labels",
".",
"append",
"(",
"label",
")",
"return",
"labels"
] | get labels . | train | false |
18,534 | def getConnectionVertexes(geometryOutput):
connectionVertexes = []
addConnectionVertexes(connectionVertexes, geometryOutput)
return connectionVertexes
| [
"def",
"getConnectionVertexes",
"(",
"geometryOutput",
")",
":",
"connectionVertexes",
"=",
"[",
"]",
"addConnectionVertexes",
"(",
"connectionVertexes",
",",
"geometryOutput",
")",
"return",
"connectionVertexes"
] | get the connections and vertexes . | train | false |
18,536 | def kstest_normal(x, pvalmethod='approx'):
x = np.asarray(x)
z = ((x - x.mean()) / x.std(ddof=1))
nobs = len(z)
d_ks = ksstat(z, stats.norm.cdf, alternative='two_sided')
if (pvalmethod == 'approx'):
pval = pval_lf(d_ks, nobs)
elif (pvalmethod == 'table'):
pval = lilliefors_table.prob(d_ks, nobs)
return (d_ks, pval)
| [
"def",
"kstest_normal",
"(",
"x",
",",
"pvalmethod",
"=",
"'approx'",
")",
":",
"x",
"=",
"np",
".",
"asarray",
"(",
"x",
")",
"z",
"=",
"(",
"(",
"x",
"-",
"x",
".",
"mean",
"(",
")",
")",
"/",
"x",
".",
"std",
"(",
"ddof",
"=",
"1",
")",
")",
"nobs",
"=",
"len",
"(",
"z",
")",
"d_ks",
"=",
"ksstat",
"(",
"z",
",",
"stats",
".",
"norm",
".",
"cdf",
",",
"alternative",
"=",
"'two_sided'",
")",
"if",
"(",
"pvalmethod",
"==",
"'approx'",
")",
":",
"pval",
"=",
"pval_lf",
"(",
"d_ks",
",",
"nobs",
")",
"elif",
"(",
"pvalmethod",
"==",
"'table'",
")",
":",
"pval",
"=",
"lilliefors_table",
".",
"prob",
"(",
"d_ks",
",",
"nobs",
")",
"return",
"(",
"d_ks",
",",
"pval",
")"
] | lilliefors test for normality . | train | false |
18,537 | def format_job_instance(job):
ret = {'Function': job.get('fun', 'unknown-function'), 'Arguments': list(job.get('arg', [])), 'Target': job.get('tgt', 'unknown-target'), 'Target-type': job.get('tgt_type', []), 'User': job.get('user', 'root')}
if ('metadata' in job):
ret['Metadata'] = job.get('metadata', {})
elif ('kwargs' in job):
if ('metadata' in job['kwargs']):
ret['Metadata'] = job['kwargs'].get('metadata', {})
return ret
| [
"def",
"format_job_instance",
"(",
"job",
")",
":",
"ret",
"=",
"{",
"'Function'",
":",
"job",
".",
"get",
"(",
"'fun'",
",",
"'unknown-function'",
")",
",",
"'Arguments'",
":",
"list",
"(",
"job",
".",
"get",
"(",
"'arg'",
",",
"[",
"]",
")",
")",
",",
"'Target'",
":",
"job",
".",
"get",
"(",
"'tgt'",
",",
"'unknown-target'",
")",
",",
"'Target-type'",
":",
"job",
".",
"get",
"(",
"'tgt_type'",
",",
"[",
"]",
")",
",",
"'User'",
":",
"job",
".",
"get",
"(",
"'user'",
",",
"'root'",
")",
"}",
"if",
"(",
"'metadata'",
"in",
"job",
")",
":",
"ret",
"[",
"'Metadata'",
"]",
"=",
"job",
".",
"get",
"(",
"'metadata'",
",",
"{",
"}",
")",
"elif",
"(",
"'kwargs'",
"in",
"job",
")",
":",
"if",
"(",
"'metadata'",
"in",
"job",
"[",
"'kwargs'",
"]",
")",
":",
"ret",
"[",
"'Metadata'",
"]",
"=",
"job",
"[",
"'kwargs'",
"]",
".",
"get",
"(",
"'metadata'",
",",
"{",
"}",
")",
"return",
"ret"
] | format the job instance correctly . | train | true |
18,538 | def _esc(code):
return '\x1b[{}m'.format(code)
| [
"def",
"_esc",
"(",
"code",
")",
":",
"return",
"'\\x1b[{}m'",
".",
"format",
"(",
"code",
")"
] | get an ansi color code based on a color number . | train | false |
18,540 | def find_language(locale):
if (not locale):
return None
LANGS = settings.AMO_LANGUAGES
if (locale in LANGS):
return locale
loc = settings.SHORTER_LANGUAGES.get(locale)
if loc:
return loc
locale = to_language(locale)
if (locale in LANGS):
return locale
return None
| [
"def",
"find_language",
"(",
"locale",
")",
":",
"if",
"(",
"not",
"locale",
")",
":",
"return",
"None",
"LANGS",
"=",
"settings",
".",
"AMO_LANGUAGES",
"if",
"(",
"locale",
"in",
"LANGS",
")",
":",
"return",
"locale",
"loc",
"=",
"settings",
".",
"SHORTER_LANGUAGES",
".",
"get",
"(",
"locale",
")",
"if",
"loc",
":",
"return",
"loc",
"locale",
"=",
"to_language",
"(",
"locale",
")",
"if",
"(",
"locale",
"in",
"LANGS",
")",
":",
"return",
"locale",
"return",
"None"
] | return a locale we support . | train | false |
18,542 | def reset_all_users():
for user in User.objects.all():
user.delete()
| [
"def",
"reset_all_users",
"(",
")",
":",
"for",
"user",
"in",
"User",
".",
"objects",
".",
"all",
"(",
")",
":",
"user",
".",
"delete",
"(",
")"
] | reset to a clean state by deleting all users . | train | false |
18,543 | def assert_logs(log_action, node_key, index=(-1)):
def outer_wrapper(func):
@functools.wraps(func)
def wrapper(self, *args, **kwargs):
node = getattr(self, node_key)
last_log = node.logs.latest()
func(self, *args, **kwargs)
node.reload()
new_log = node.logs.order_by('-date')[((- index) - 1)]
assert_not_equal(last_log._id, new_log._id)
assert_equal(new_log.action, log_action)
node.save()
return wrapper
return outer_wrapper
| [
"def",
"assert_logs",
"(",
"log_action",
",",
"node_key",
",",
"index",
"=",
"(",
"-",
"1",
")",
")",
":",
"def",
"outer_wrapper",
"(",
"func",
")",
":",
"@",
"functools",
".",
"wraps",
"(",
"func",
")",
"def",
"wrapper",
"(",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"node",
"=",
"getattr",
"(",
"self",
",",
"node_key",
")",
"last_log",
"=",
"node",
".",
"logs",
".",
"latest",
"(",
")",
"func",
"(",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"node",
".",
"reload",
"(",
")",
"new_log",
"=",
"node",
".",
"logs",
".",
"order_by",
"(",
"'-date'",
")",
"[",
"(",
"(",
"-",
"index",
")",
"-",
"1",
")",
"]",
"assert_not_equal",
"(",
"last_log",
".",
"_id",
",",
"new_log",
".",
"_id",
")",
"assert_equal",
"(",
"new_log",
".",
"action",
",",
"log_action",
")",
"node",
".",
"save",
"(",
")",
"return",
"wrapper",
"return",
"outer_wrapper"
] | a decorator to ensure a log is added during a unit test . | train | false |
18,544 | def _verify_dict_keys(expected_keys, target_dict, strict=True):
if (not isinstance(target_dict, dict)):
msg = (_("Invalid input. '%(target_dict)s' must be a dictionary with keys: %(expected_keys)s") % dict(target_dict=target_dict, expected_keys=expected_keys))
return msg
expected_keys = set(expected_keys)
provided_keys = set(target_dict.keys())
predicate = (expected_keys.__eq__ if strict else expected_keys.issubset)
if (not predicate(provided_keys)):
msg = (_("Validation of dictionary's keys failed.Expected keys: %(expected_keys)s Provided keys: %(provided_keys)s") % locals())
return msg
| [
"def",
"_verify_dict_keys",
"(",
"expected_keys",
",",
"target_dict",
",",
"strict",
"=",
"True",
")",
":",
"if",
"(",
"not",
"isinstance",
"(",
"target_dict",
",",
"dict",
")",
")",
":",
"msg",
"=",
"(",
"_",
"(",
"\"Invalid input. '%(target_dict)s' must be a dictionary with keys: %(expected_keys)s\"",
")",
"%",
"dict",
"(",
"target_dict",
"=",
"target_dict",
",",
"expected_keys",
"=",
"expected_keys",
")",
")",
"return",
"msg",
"expected_keys",
"=",
"set",
"(",
"expected_keys",
")",
"provided_keys",
"=",
"set",
"(",
"target_dict",
".",
"keys",
"(",
")",
")",
"predicate",
"=",
"(",
"expected_keys",
".",
"__eq__",
"if",
"strict",
"else",
"expected_keys",
".",
"issubset",
")",
"if",
"(",
"not",
"predicate",
"(",
"provided_keys",
")",
")",
":",
"msg",
"=",
"(",
"_",
"(",
"\"Validation of dictionary's keys failed.Expected keys: %(expected_keys)s Provided keys: %(provided_keys)s\"",
")",
"%",
"locals",
"(",
")",
")",
"return",
"msg"
] | allows to verify keys in a dictionary . | train | false |
18,546 | def calc_expected_status_length(status, short_url_length=23):
status_length = 0
for word in re.split(u'\\s', status):
if is_url(word):
status_length += short_url_length
else:
status_length += len(word)
status_length += len(re.findall(u'\\s', status))
return status_length
| [
"def",
"calc_expected_status_length",
"(",
"status",
",",
"short_url_length",
"=",
"23",
")",
":",
"status_length",
"=",
"0",
"for",
"word",
"in",
"re",
".",
"split",
"(",
"u'\\\\s'",
",",
"status",
")",
":",
"if",
"is_url",
"(",
"word",
")",
":",
"status_length",
"+=",
"short_url_length",
"else",
":",
"status_length",
"+=",
"len",
"(",
"word",
")",
"status_length",
"+=",
"len",
"(",
"re",
".",
"findall",
"(",
"u'\\\\s'",
",",
"status",
")",
")",
"return",
"status_length"
] | calculates the length of a tweet . | train | false |
18,547 | def tuplize_dict(data_dict):
tuplized_dict = {}
for (key, value) in data_dict.iteritems():
key_list = key.split('__')
for (num, key) in enumerate(key_list):
if ((num % 2) == 1):
try:
key_list[num] = int(key)
except ValueError:
raise df.DataError('Bad key')
tuplized_dict[tuple(key_list)] = value
return tuplized_dict
| [
"def",
"tuplize_dict",
"(",
"data_dict",
")",
":",
"tuplized_dict",
"=",
"{",
"}",
"for",
"(",
"key",
",",
"value",
")",
"in",
"data_dict",
".",
"iteritems",
"(",
")",
":",
"key_list",
"=",
"key",
".",
"split",
"(",
"'__'",
")",
"for",
"(",
"num",
",",
"key",
")",
"in",
"enumerate",
"(",
"key_list",
")",
":",
"if",
"(",
"(",
"num",
"%",
"2",
")",
"==",
"1",
")",
":",
"try",
":",
"key_list",
"[",
"num",
"]",
"=",
"int",
"(",
"key",
")",
"except",
"ValueError",
":",
"raise",
"df",
".",
"DataError",
"(",
"'Bad key'",
")",
"tuplized_dict",
"[",
"tuple",
"(",
"key_list",
")",
"]",
"=",
"value",
"return",
"tuplized_dict"
] | takes a dict with keys of the form table__0__key and converts them to a tuple like . | train | false |
18,548 | def elide(text, length):
if (length < 1):
raise ValueError('length must be >= 1!')
if (len(text) <= length):
return text
else:
return (text[:(length - 1)] + '\\u2026')
| [
"def",
"elide",
"(",
"text",
",",
"length",
")",
":",
"if",
"(",
"length",
"<",
"1",
")",
":",
"raise",
"ValueError",
"(",
"'length must be >= 1!'",
")",
"if",
"(",
"len",
"(",
"text",
")",
"<=",
"length",
")",
":",
"return",
"text",
"else",
":",
"return",
"(",
"text",
"[",
":",
"(",
"length",
"-",
"1",
")",
"]",
"+",
"'\\\\u2026'",
")"
] | elide text so it uses a maximum of length chars . | train | false |
18,549 | def generate_data(n_subjects, n_conditions):
rng = np.random.RandomState(42)
data = rng.randn((n_subjects * n_conditions)).reshape(n_subjects, n_conditions)
return data
| [
"def",
"generate_data",
"(",
"n_subjects",
",",
"n_conditions",
")",
":",
"rng",
"=",
"np",
".",
"random",
".",
"RandomState",
"(",
"42",
")",
"data",
"=",
"rng",
".",
"randn",
"(",
"(",
"n_subjects",
"*",
"n_conditions",
")",
")",
".",
"reshape",
"(",
"n_subjects",
",",
"n_conditions",
")",
"return",
"data"
] | generate random blob-ish data with noisy features . | train | false |
18,550 | def change_DAILYSEARCH_FREQUENCY(freq):
sickbeard.DAILYSEARCH_FREQUENCY = try_int(freq, sickbeard.DEFAULT_DAILYSEARCH_FREQUENCY)
if (sickbeard.DAILYSEARCH_FREQUENCY < sickbeard.MIN_DAILYSEARCH_FREQUENCY):
sickbeard.DAILYSEARCH_FREQUENCY = sickbeard.MIN_DAILYSEARCH_FREQUENCY
sickbeard.dailySearchScheduler.cycleTime = datetime.timedelta(minutes=sickbeard.DAILYSEARCH_FREQUENCY)
| [
"def",
"change_DAILYSEARCH_FREQUENCY",
"(",
"freq",
")",
":",
"sickbeard",
".",
"DAILYSEARCH_FREQUENCY",
"=",
"try_int",
"(",
"freq",
",",
"sickbeard",
".",
"DEFAULT_DAILYSEARCH_FREQUENCY",
")",
"if",
"(",
"sickbeard",
".",
"DAILYSEARCH_FREQUENCY",
"<",
"sickbeard",
".",
"MIN_DAILYSEARCH_FREQUENCY",
")",
":",
"sickbeard",
".",
"DAILYSEARCH_FREQUENCY",
"=",
"sickbeard",
".",
"MIN_DAILYSEARCH_FREQUENCY",
"sickbeard",
".",
"dailySearchScheduler",
".",
"cycleTime",
"=",
"datetime",
".",
"timedelta",
"(",
"minutes",
"=",
"sickbeard",
".",
"DAILYSEARCH_FREQUENCY",
")"
] | change frequency of daily search thread . | train | false |
18,551 | def general_gaussian(M, p, sig, sym=True):
if _len_guards(M):
return np.ones(M)
(M, needs_trunc) = _extend(M, sym)
n = (np.arange(0, M) - ((M - 1.0) / 2.0))
w = np.exp(((-0.5) * (np.abs((n / sig)) ** (2 * p))))
return _truncate(w, needs_trunc)
| [
"def",
"general_gaussian",
"(",
"M",
",",
"p",
",",
"sig",
",",
"sym",
"=",
"True",
")",
":",
"if",
"_len_guards",
"(",
"M",
")",
":",
"return",
"np",
".",
"ones",
"(",
"M",
")",
"(",
"M",
",",
"needs_trunc",
")",
"=",
"_extend",
"(",
"M",
",",
"sym",
")",
"n",
"=",
"(",
"np",
".",
"arange",
"(",
"0",
",",
"M",
")",
"-",
"(",
"(",
"M",
"-",
"1.0",
")",
"/",
"2.0",
")",
")",
"w",
"=",
"np",
".",
"exp",
"(",
"(",
"(",
"-",
"0.5",
")",
"*",
"(",
"np",
".",
"abs",
"(",
"(",
"n",
"/",
"sig",
")",
")",
"**",
"(",
"2",
"*",
"p",
")",
")",
")",
")",
"return",
"_truncate",
"(",
"w",
",",
"needs_trunc",
")"
] | return a window with a generalized gaussian shape . | train | false |
18,552 | def mean_squared_error(y_true, y_pred):
return tf.reduce_mean(tf.square((y_pred - y_true)))
| [
"def",
"mean_squared_error",
"(",
"y_true",
",",
"y_pred",
")",
":",
"return",
"tf",
".",
"reduce_mean",
"(",
"tf",
".",
"square",
"(",
"(",
"y_pred",
"-",
"y_true",
")",
")",
")"
] | l2 distance between tensors true and pred . | train | false |
18,553 | def test_angle_format_roundtripping():
a1 = Angle(0, unit=u.radian)
a2 = Angle(10, unit=u.degree)
a3 = Angle(0.543, unit=u.degree)
a4 = Angle(u'1d2m3.4s')
assert (Angle(str(a1)).degree == a1.degree)
assert (Angle(str(a2)).degree == a2.degree)
assert (Angle(str(a3)).degree == a3.degree)
assert (Angle(str(a4)).degree == a4.degree)
ra = Longitude(u'1h2m3.4s')
dec = Latitude(u'1d2m3.4s')
assert_allclose(Angle(str(ra)).degree, ra.degree)
assert_allclose(Angle(str(dec)).degree, dec.degree)
| [
"def",
"test_angle_format_roundtripping",
"(",
")",
":",
"a1",
"=",
"Angle",
"(",
"0",
",",
"unit",
"=",
"u",
".",
"radian",
")",
"a2",
"=",
"Angle",
"(",
"10",
",",
"unit",
"=",
"u",
".",
"degree",
")",
"a3",
"=",
"Angle",
"(",
"0.543",
",",
"unit",
"=",
"u",
".",
"degree",
")",
"a4",
"=",
"Angle",
"(",
"u'1d2m3.4s'",
")",
"assert",
"(",
"Angle",
"(",
"str",
"(",
"a1",
")",
")",
".",
"degree",
"==",
"a1",
".",
"degree",
")",
"assert",
"(",
"Angle",
"(",
"str",
"(",
"a2",
")",
")",
".",
"degree",
"==",
"a2",
".",
"degree",
")",
"assert",
"(",
"Angle",
"(",
"str",
"(",
"a3",
")",
")",
".",
"degree",
"==",
"a3",
".",
"degree",
")",
"assert",
"(",
"Angle",
"(",
"str",
"(",
"a4",
")",
")",
".",
"degree",
"==",
"a4",
".",
"degree",
")",
"ra",
"=",
"Longitude",
"(",
"u'1h2m3.4s'",
")",
"dec",
"=",
"Latitude",
"(",
"u'1d2m3.4s'",
")",
"assert_allclose",
"(",
"Angle",
"(",
"str",
"(",
"ra",
")",
")",
".",
"degree",
",",
"ra",
".",
"degree",
")",
"assert_allclose",
"(",
"Angle",
"(",
"str",
"(",
"dec",
")",
")",
".",
"degree",
",",
"dec",
".",
"degree",
")"
] | ensures that the string representation of an angle can be used to create a new valid angle . | train | false |
18,554 | def _comment_line(line):
line = line.rstrip()
if line:
return ('# ' + line)
else:
return '#'
| [
"def",
"_comment_line",
"(",
"line",
")",
":",
"line",
"=",
"line",
".",
"rstrip",
"(",
")",
"if",
"line",
":",
"return",
"(",
"'# '",
"+",
"line",
")",
"else",
":",
"return",
"'#'"
] | return a commented form of the given line . | train | false |
18,556 | def list_agents(profile=None):
conn = _auth(profile)
return conn.list_agents()
| [
"def",
"list_agents",
"(",
"profile",
"=",
"None",
")",
":",
"conn",
"=",
"_auth",
"(",
"profile",
")",
"return",
"conn",
".",
"list_agents",
"(",
")"
] | list agents . | train | false |
18,558 | def list_empty(strings):
for string in strings:
if (len(string) > 0):
return False
return True
| [
"def",
"list_empty",
"(",
"strings",
")",
":",
"for",
"string",
"in",
"strings",
":",
"if",
"(",
"len",
"(",
"string",
")",
">",
"0",
")",
":",
"return",
"False",
"return",
"True"
] | check if list is exclusively made of empty strings . | train | false |
18,559 | def device_to_device(dst, src, size, stream=0):
varargs = []
if stream:
assert isinstance(stream, Stream)
fn = driver.cuMemcpyDtoDAsync
varargs.append(stream.handle)
else:
fn = driver.cuMemcpyDtoD
fn(device_pointer(dst), device_pointer(src), size, *varargs)
| [
"def",
"device_to_device",
"(",
"dst",
",",
"src",
",",
"size",
",",
"stream",
"=",
"0",
")",
":",
"varargs",
"=",
"[",
"]",
"if",
"stream",
":",
"assert",
"isinstance",
"(",
"stream",
",",
"Stream",
")",
"fn",
"=",
"driver",
".",
"cuMemcpyDtoDAsync",
"varargs",
".",
"append",
"(",
"stream",
".",
"handle",
")",
"else",
":",
"fn",
"=",
"driver",
".",
"cuMemcpyDtoD",
"fn",
"(",
"device_pointer",
"(",
"dst",
")",
",",
"device_pointer",
"(",
"src",
")",
",",
"size",
",",
"*",
"varargs",
")"
] | note: the underlying data pointer from the host data buffer is used and it should not be changed until the operation which can be asynchronous completes . | train | false |
18,560 | def vector_product(v0, v1, axis=0):
return numpy.cross(v0, v1, axis=axis)
| [
"def",
"vector_product",
"(",
"v0",
",",
"v1",
",",
"axis",
"=",
"0",
")",
":",
"return",
"numpy",
".",
"cross",
"(",
"v0",
",",
"v1",
",",
"axis",
"=",
"axis",
")"
] | return vector perpendicular to vectors . | train | true |
18,562 | def guvectorize(ftylist, signature, **kws):
if isinstance(ftylist, str):
ftylist = [ftylist]
def wrap(func):
guvec = GUVectorize(func, signature, **kws)
for fty in ftylist:
guvec.add(fty)
return guvec.build_ufunc()
return wrap
| [
"def",
"guvectorize",
"(",
"ftylist",
",",
"signature",
",",
"**",
"kws",
")",
":",
"if",
"isinstance",
"(",
"ftylist",
",",
"str",
")",
":",
"ftylist",
"=",
"[",
"ftylist",
"]",
"def",
"wrap",
"(",
"func",
")",
":",
"guvec",
"=",
"GUVectorize",
"(",
"func",
",",
"signature",
",",
"**",
"kws",
")",
"for",
"fty",
"in",
"ftylist",
":",
"guvec",
".",
"add",
"(",
"fty",
")",
"return",
"guvec",
".",
"build_ufunc",
"(",
")",
"return",
"wrap"
] | guvectorize a decorator to create numpy generialized-ufunc object from numba compiled code . | train | false |
18,563 | def show_source(request, project, subproject):
(obj, source) = get_source(request, project, subproject)
return render(request, 'source.html', {'object': obj, 'project': obj.project, 'source': source, 'title': (_('Source strings in %s') % force_text(obj))})
| [
"def",
"show_source",
"(",
"request",
",",
"project",
",",
"subproject",
")",
":",
"(",
"obj",
",",
"source",
")",
"=",
"get_source",
"(",
"request",
",",
"project",
",",
"subproject",
")",
"return",
"render",
"(",
"request",
",",
"'source.html'",
",",
"{",
"'object'",
":",
"obj",
",",
"'project'",
":",
"obj",
".",
"project",
",",
"'source'",
":",
"source",
",",
"'title'",
":",
"(",
"_",
"(",
"'Source strings in %s'",
")",
"%",
"force_text",
"(",
"obj",
")",
")",
"}",
")"
] | show source strings summary and checks . | train | false |
18,564 | def handleCheckIDRequest(request, openid_request):
if (not openid_request.idSelect()):
id_url = getViewURL(request, idPage)
if (id_url != openid_request.identity):
error_response = ProtocolError(openid_request.message, ('This server cannot verify the URL %r' % (openid_request.identity,)))
return displayResponse(request, error_response)
if openid_request.immediate:
openid_response = openid_request.answer(False)
return displayResponse(request, openid_response)
else:
setRequest(request, openid_request)
return showDecidePage(request, openid_request)
| [
"def",
"handleCheckIDRequest",
"(",
"request",
",",
"openid_request",
")",
":",
"if",
"(",
"not",
"openid_request",
".",
"idSelect",
"(",
")",
")",
":",
"id_url",
"=",
"getViewURL",
"(",
"request",
",",
"idPage",
")",
"if",
"(",
"id_url",
"!=",
"openid_request",
".",
"identity",
")",
":",
"error_response",
"=",
"ProtocolError",
"(",
"openid_request",
".",
"message",
",",
"(",
"'This server cannot verify the URL %r'",
"%",
"(",
"openid_request",
".",
"identity",
",",
")",
")",
")",
"return",
"displayResponse",
"(",
"request",
",",
"error_response",
")",
"if",
"openid_request",
".",
"immediate",
":",
"openid_response",
"=",
"openid_request",
".",
"answer",
"(",
"False",
")",
"return",
"displayResponse",
"(",
"request",
",",
"openid_response",
")",
"else",
":",
"setRequest",
"(",
"request",
",",
"openid_request",
")",
"return",
"showDecidePage",
"(",
"request",
",",
"openid_request",
")"
] | handle checkid_* requests . | train | true |
18,565 | def version_dict(version):
match = version_re.match((version or ''))
letters = 'alpha pre'.split()
numbers = 'major minor1 minor2 minor3 alpha_ver pre_ver'.split()
if match:
d = match.groupdict()
for letter in letters:
d[letter] = (d[letter] if d[letter] else None)
for num in numbers:
if (d[num] == '*'):
d[num] = 99
else:
d[num] = (int(d[num]) if d[num] else None)
else:
d = dict(((k, None) for k in numbers))
d.update(((k, None) for k in letters))
return d
| [
"def",
"version_dict",
"(",
"version",
")",
":",
"match",
"=",
"version_re",
".",
"match",
"(",
"(",
"version",
"or",
"''",
")",
")",
"letters",
"=",
"'alpha pre'",
".",
"split",
"(",
")",
"numbers",
"=",
"'major minor1 minor2 minor3 alpha_ver pre_ver'",
".",
"split",
"(",
")",
"if",
"match",
":",
"d",
"=",
"match",
".",
"groupdict",
"(",
")",
"for",
"letter",
"in",
"letters",
":",
"d",
"[",
"letter",
"]",
"=",
"(",
"d",
"[",
"letter",
"]",
"if",
"d",
"[",
"letter",
"]",
"else",
"None",
")",
"for",
"num",
"in",
"numbers",
":",
"if",
"(",
"d",
"[",
"num",
"]",
"==",
"'*'",
")",
":",
"d",
"[",
"num",
"]",
"=",
"99",
"else",
":",
"d",
"[",
"num",
"]",
"=",
"(",
"int",
"(",
"d",
"[",
"num",
"]",
")",
"if",
"d",
"[",
"num",
"]",
"else",
"None",
")",
"else",
":",
"d",
"=",
"dict",
"(",
"(",
"(",
"k",
",",
"None",
")",
"for",
"k",
"in",
"numbers",
")",
")",
"d",
".",
"update",
"(",
"(",
"(",
"k",
",",
"None",
")",
"for",
"k",
"in",
"letters",
")",
")",
"return",
"d"
] | turn a version string into a dict with major/minor/ . | train | true |
18,566 | def encode_local(string):
if is_python3():
return string
return string.encode((sys.getfilesystemencoding() or 'utf-8'))
| [
"def",
"encode_local",
"(",
"string",
")",
":",
"if",
"is_python3",
"(",
")",
":",
"return",
"string",
"return",
"string",
".",
"encode",
"(",
"(",
"sys",
".",
"getfilesystemencoding",
"(",
")",
"or",
"'utf-8'",
")",
")"
] | converts string into users preferred encoding . | train | false |
18,568 | def add_s3(command_table, session, **kwargs):
utils.rename_command(command_table, 's3', 's3api')
command_table['s3'] = S3(session)
| [
"def",
"add_s3",
"(",
"command_table",
",",
"session",
",",
"**",
"kwargs",
")",
":",
"utils",
".",
"rename_command",
"(",
"command_table",
",",
"'s3'",
",",
"'s3api'",
")",
"command_table",
"[",
"'s3'",
"]",
"=",
"S3",
"(",
"session",
")"
] | this creates a new service object for the s3 plugin . | train | false |
18,571 | def get_file_systems(filesystemid=None, keyid=None, key=None, profile=None, region=None, **kwargs):
result = None
client = _get_conn(key=key, keyid=keyid, profile=profile, region=region)
if filesystemid:
response = client.describe_file_systems(FileSystemId=filesystemid)
result = response['FileSystems']
else:
response = client.describe_file_systems()
result = response['FileSystems']
while ('NextMarker' in response):
response = client.describe_file_systems(Marker=response['NextMarker'])
result.extend(response['FileSystems'])
return result
| [
"def",
"get_file_systems",
"(",
"filesystemid",
"=",
"None",
",",
"keyid",
"=",
"None",
",",
"key",
"=",
"None",
",",
"profile",
"=",
"None",
",",
"region",
"=",
"None",
",",
"**",
"kwargs",
")",
":",
"result",
"=",
"None",
"client",
"=",
"_get_conn",
"(",
"key",
"=",
"key",
",",
"keyid",
"=",
"keyid",
",",
"profile",
"=",
"profile",
",",
"region",
"=",
"region",
")",
"if",
"filesystemid",
":",
"response",
"=",
"client",
".",
"describe_file_systems",
"(",
"FileSystemId",
"=",
"filesystemid",
")",
"result",
"=",
"response",
"[",
"'FileSystems'",
"]",
"else",
":",
"response",
"=",
"client",
".",
"describe_file_systems",
"(",
")",
"result",
"=",
"response",
"[",
"'FileSystems'",
"]",
"while",
"(",
"'NextMarker'",
"in",
"response",
")",
":",
"response",
"=",
"client",
".",
"describe_file_systems",
"(",
"Marker",
"=",
"response",
"[",
"'NextMarker'",
"]",
")",
"result",
".",
"extend",
"(",
"response",
"[",
"'FileSystems'",
"]",
")",
"return",
"result"
] | get all efs properties or a specific instance property if filesystemid is specified filesystemid - id of the file system to retrieve properties returns - list of all elastic file system properties cli example: . | train | false |
18,572 | def json_formatall(filenames, indent=DEFAULT_INDENT_SIZE, dry_run=False):
errors = 0
console = logging.getLogger('console')
for filename in filenames:
try:
result = json_format(filename, indent=indent, console=console, dry_run=dry_run)
if (not result):
errors += 1
except Exception as e:
console.error('ERROR %s: %s (filename: %s)', e.__class__.__name__, e, filename)
errors += 1
return errors
| [
"def",
"json_formatall",
"(",
"filenames",
",",
"indent",
"=",
"DEFAULT_INDENT_SIZE",
",",
"dry_run",
"=",
"False",
")",
":",
"errors",
"=",
"0",
"console",
"=",
"logging",
".",
"getLogger",
"(",
"'console'",
")",
"for",
"filename",
"in",
"filenames",
":",
"try",
":",
"result",
"=",
"json_format",
"(",
"filename",
",",
"indent",
"=",
"indent",
",",
"console",
"=",
"console",
",",
"dry_run",
"=",
"dry_run",
")",
"if",
"(",
"not",
"result",
")",
":",
"errors",
"+=",
"1",
"except",
"Exception",
"as",
"e",
":",
"console",
".",
"error",
"(",
"'ERROR %s: %s (filename: %s)'",
",",
"e",
".",
"__class__",
".",
"__name__",
",",
"e",
",",
"filename",
")",
"errors",
"+=",
"1",
"return",
"errors"
] | format/beautify a json file . | train | false |
18,573 | def ImportAndCall(node, results, names):
obj = results['obj'].clone()
if (obj.type == syms.arglist):
newarglist = obj.clone()
else:
newarglist = Node(syms.arglist, [obj.clone()])
after = results['after']
if after:
after = [n.clone() for n in after]
new = Node(syms.power, ((Attr(Name(names[0]), Name(names[1])) + [Node(syms.trailer, [results['lpar'].clone(), newarglist, results['rpar'].clone()])]) + after))
new.prefix = node.prefix
return new
| [
"def",
"ImportAndCall",
"(",
"node",
",",
"results",
",",
"names",
")",
":",
"obj",
"=",
"results",
"[",
"'obj'",
"]",
".",
"clone",
"(",
")",
"if",
"(",
"obj",
".",
"type",
"==",
"syms",
".",
"arglist",
")",
":",
"newarglist",
"=",
"obj",
".",
"clone",
"(",
")",
"else",
":",
"newarglist",
"=",
"Node",
"(",
"syms",
".",
"arglist",
",",
"[",
"obj",
".",
"clone",
"(",
")",
"]",
")",
"after",
"=",
"results",
"[",
"'after'",
"]",
"if",
"after",
":",
"after",
"=",
"[",
"n",
".",
"clone",
"(",
")",
"for",
"n",
"in",
"after",
"]",
"new",
"=",
"Node",
"(",
"syms",
".",
"power",
",",
"(",
"(",
"Attr",
"(",
"Name",
"(",
"names",
"[",
"0",
"]",
")",
",",
"Name",
"(",
"names",
"[",
"1",
"]",
")",
")",
"+",
"[",
"Node",
"(",
"syms",
".",
"trailer",
",",
"[",
"results",
"[",
"'lpar'",
"]",
".",
"clone",
"(",
")",
",",
"newarglist",
",",
"results",
"[",
"'rpar'",
"]",
".",
"clone",
"(",
")",
"]",
")",
"]",
")",
"+",
"after",
")",
")",
"new",
".",
"prefix",
"=",
"node",
".",
"prefix",
"return",
"new"
] | returns an import statement and calls a method of the module: import module module . | train | false |
18,574 | def test_lessthan(value, other):
return (value < other)
| [
"def",
"test_lessthan",
"(",
"value",
",",
"other",
")",
":",
"return",
"(",
"value",
"<",
"other",
")"
] | check if value is less than other . | train | false |
18,575 | def key_type(key, host=None, port=None, db=None, password=None):
server = _connect(host, port, db, password)
return server.type(key)
| [
"def",
"key_type",
"(",
"key",
",",
"host",
"=",
"None",
",",
"port",
"=",
"None",
",",
"db",
"=",
"None",
",",
"password",
"=",
"None",
")",
":",
"server",
"=",
"_connect",
"(",
"host",
",",
"port",
",",
"db",
",",
"password",
")",
"return",
"server",
".",
"type",
"(",
"key",
")"
] | get redis key type cli example: . | train | true |
18,577 | def is_redirection(status):
return (300 <= status <= 399)
| [
"def",
"is_redirection",
"(",
"status",
")",
":",
"return",
"(",
"300",
"<=",
"status",
"<=",
"399",
")"
] | check if http status code is redirection . | train | false |
18,578 | def add_model_for_resource(resource, model):
_RESOURCE_TO_MODEL_MAP[resource] = model
| [
"def",
"add_model_for_resource",
"(",
"resource",
",",
"model",
")",
":",
"_RESOURCE_TO_MODEL_MAP",
"[",
"resource",
"]",
"=",
"model"
] | adds a mapping between a callback resource and a db model . | train | false |
18,580 | def fmt_whitespace(value):
value = WHITESPACE_RE.sub(u'<span class="hlspace">\\1</span>', value)
value = value.replace(u' DCTB ', SPACE_TAB.format(_(u'Tab character')))
return value
| [
"def",
"fmt_whitespace",
"(",
"value",
")",
":",
"value",
"=",
"WHITESPACE_RE",
".",
"sub",
"(",
"u'<span class=\"hlspace\">\\\\1</span>'",
",",
"value",
")",
"value",
"=",
"value",
".",
"replace",
"(",
"u' DCTB '",
",",
"SPACE_TAB",
".",
"format",
"(",
"_",
"(",
"u'Tab character'",
")",
")",
")",
"return",
"value"
] | formats whitespace so that it is more visible . | train | false |
18,581 | @task(aliases=['pack'])
def webpack(ctx, clean=False, watch=False, dev=False):
if clean:
clean_assets(ctx)
if (os.getcwd() != HERE):
os.chdir(HERE)
webpack_bin = os.path.join(HERE, 'node_modules', 'webpack', 'bin', 'webpack.js')
args = [webpack_bin]
if (settings.DEBUG_MODE and dev):
args += ['--colors']
else:
args += ['--progress']
if watch:
args += ['--watch']
config_file = ('webpack.admin.config.js' if dev else 'webpack.prod.config.js')
args += ['--config {0}'.format(config_file)]
command = ' '.join(args)
ctx.run(command, echo=True)
| [
"@",
"task",
"(",
"aliases",
"=",
"[",
"'pack'",
"]",
")",
"def",
"webpack",
"(",
"ctx",
",",
"clean",
"=",
"False",
",",
"watch",
"=",
"False",
",",
"dev",
"=",
"False",
")",
":",
"if",
"clean",
":",
"clean_assets",
"(",
"ctx",
")",
"if",
"(",
"os",
".",
"getcwd",
"(",
")",
"!=",
"HERE",
")",
":",
"os",
".",
"chdir",
"(",
"HERE",
")",
"webpack_bin",
"=",
"os",
".",
"path",
".",
"join",
"(",
"HERE",
",",
"'node_modules'",
",",
"'webpack'",
",",
"'bin'",
",",
"'webpack.js'",
")",
"args",
"=",
"[",
"webpack_bin",
"]",
"if",
"(",
"settings",
".",
"DEBUG_MODE",
"and",
"dev",
")",
":",
"args",
"+=",
"[",
"'--colors'",
"]",
"else",
":",
"args",
"+=",
"[",
"'--progress'",
"]",
"if",
"watch",
":",
"args",
"+=",
"[",
"'--watch'",
"]",
"config_file",
"=",
"(",
"'webpack.admin.config.js'",
"if",
"dev",
"else",
"'webpack.prod.config.js'",
")",
"args",
"+=",
"[",
"'--config {0}'",
".",
"format",
"(",
"config_file",
")",
"]",
"command",
"=",
"' '",
".",
"join",
"(",
"args",
")",
"ctx",
".",
"run",
"(",
"command",
",",
"echo",
"=",
"True",
")"
] | build static assets with webpack . | train | false |
18,582 | @contextmanager
def in_dir(dir=None):
cwd = os.getcwd()
if (dir is None):
(yield cwd)
return
os.chdir(dir)
(yield dir)
os.chdir(cwd)
| [
"@",
"contextmanager",
"def",
"in_dir",
"(",
"dir",
"=",
"None",
")",
":",
"cwd",
"=",
"os",
".",
"getcwd",
"(",
")",
"if",
"(",
"dir",
"is",
"None",
")",
":",
"(",
"yield",
"cwd",
")",
"return",
"os",
".",
"chdir",
"(",
"dir",
")",
"(",
"yield",
"dir",
")",
"os",
".",
"chdir",
"(",
"cwd",
")"
] | change directory to given directory for duration of with block useful when you want to use in_tempdir for the final test . | train | false |
18,583 | def upgrade_master(ansible_module, config_base, from_version, to_version, backup):
if (from_version == '3.0'):
if (to_version == '3.1'):
return upgrade_master_3_0_to_3_1(ansible_module, config_base, backup)
| [
"def",
"upgrade_master",
"(",
"ansible_module",
",",
"config_base",
",",
"from_version",
",",
"to_version",
",",
"backup",
")",
":",
"if",
"(",
"from_version",
"==",
"'3.0'",
")",
":",
"if",
"(",
"to_version",
"==",
"'3.1'",
")",
":",
"return",
"upgrade_master_3_0_to_3_1",
"(",
"ansible_module",
",",
"config_base",
",",
"backup",
")"
] | upgrade entry point . | train | false |
18,584 | def connect_to_cloud_loadbalancers(region=None):
return _create_client(ep_name='load_balancer', region=region)
| [
"def",
"connect_to_cloud_loadbalancers",
"(",
"region",
"=",
"None",
")",
":",
"return",
"_create_client",
"(",
"ep_name",
"=",
"'load_balancer'",
",",
"region",
"=",
"region",
")"
] | creates a client for working with cloud loadbalancers . | train | false |
18,585 | def doctest_depends_on(exe=None, modules=None, disable_viewers=None):
pyglet = False
if ((modules is not None) and ('pyglet' in modules)):
pyglet = True
def depends_on_deco(fn):
fn._doctest_depends_on = dict(exe=exe, modules=modules, disable_viewers=disable_viewers, pyglet=pyglet)
if inspect.isclass(fn):
fn._doctest_depdends_on = no_attrs_in_subclass(fn, fn._doctest_depends_on)
return fn
return depends_on_deco
| [
"def",
"doctest_depends_on",
"(",
"exe",
"=",
"None",
",",
"modules",
"=",
"None",
",",
"disable_viewers",
"=",
"None",
")",
":",
"pyglet",
"=",
"False",
"if",
"(",
"(",
"modules",
"is",
"not",
"None",
")",
"and",
"(",
"'pyglet'",
"in",
"modules",
")",
")",
":",
"pyglet",
"=",
"True",
"def",
"depends_on_deco",
"(",
"fn",
")",
":",
"fn",
".",
"_doctest_depends_on",
"=",
"dict",
"(",
"exe",
"=",
"exe",
",",
"modules",
"=",
"modules",
",",
"disable_viewers",
"=",
"disable_viewers",
",",
"pyglet",
"=",
"pyglet",
")",
"if",
"inspect",
".",
"isclass",
"(",
"fn",
")",
":",
"fn",
".",
"_doctest_depdends_on",
"=",
"no_attrs_in_subclass",
"(",
"fn",
",",
"fn",
".",
"_doctest_depends_on",
")",
"return",
"fn",
"return",
"depends_on_deco"
] | adds metadata about the depenencies which need to be met for doctesting the docstrings of the decorated objects . | train | false |
18,587 | def GetFirst(parameters, key, default=None):
if (key in parameters):
if parameters[key]:
return parameters[key][0]
return default
| [
"def",
"GetFirst",
"(",
"parameters",
",",
"key",
",",
"default",
"=",
"None",
")",
":",
"if",
"(",
"key",
"in",
"parameters",
")",
":",
"if",
"parameters",
"[",
"key",
"]",
":",
"return",
"parameters",
"[",
"key",
"]",
"[",
"0",
"]",
"return",
"default"
] | returns the first value of the given key . | train | false |
18,588 | def is_number_match(num1, num2):
if (isinstance(num1, PhoneNumber) and isinstance(num2, PhoneNumber)):
return _is_number_match_OO(num1, num2)
elif isinstance(num1, PhoneNumber):
return _is_number_match_OS(num1, num2)
elif isinstance(num2, PhoneNumber):
return _is_number_match_OS(num2, num1)
else:
return _is_number_match_SS(num1, num2)
| [
"def",
"is_number_match",
"(",
"num1",
",",
"num2",
")",
":",
"if",
"(",
"isinstance",
"(",
"num1",
",",
"PhoneNumber",
")",
"and",
"isinstance",
"(",
"num2",
",",
"PhoneNumber",
")",
")",
":",
"return",
"_is_number_match_OO",
"(",
"num1",
",",
"num2",
")",
"elif",
"isinstance",
"(",
"num1",
",",
"PhoneNumber",
")",
":",
"return",
"_is_number_match_OS",
"(",
"num1",
",",
"num2",
")",
"elif",
"isinstance",
"(",
"num2",
",",
"PhoneNumber",
")",
":",
"return",
"_is_number_match_OS",
"(",
"num2",
",",
"num1",
")",
"else",
":",
"return",
"_is_number_match_SS",
"(",
"num1",
",",
"num2",
")"
] | takes two phone numbers and compares them for equality . | train | true |
18,590 | def write_delete_marker_file(directory):
filepath = os.path.join(directory, PIP_DELETE_MARKER_FILENAME)
marker_fp = open(filepath, 'w')
marker_fp.write(DELETE_MARKER_MESSAGE)
marker_fp.close()
| [
"def",
"write_delete_marker_file",
"(",
"directory",
")",
":",
"filepath",
"=",
"os",
".",
"path",
".",
"join",
"(",
"directory",
",",
"PIP_DELETE_MARKER_FILENAME",
")",
"marker_fp",
"=",
"open",
"(",
"filepath",
",",
"'w'",
")",
"marker_fp",
".",
"write",
"(",
"DELETE_MARKER_MESSAGE",
")",
"marker_fp",
".",
"close",
"(",
")"
] | write the pip delete marker file into this directory . | train | true |
18,594 | def hashvalue(arg):
if isinstance(arg, XBlock):
return unicode(arg.location)
else:
return unicode(arg)
| [
"def",
"hashvalue",
"(",
"arg",
")",
":",
"if",
"isinstance",
"(",
"arg",
",",
"XBlock",
")",
":",
"return",
"unicode",
"(",
"arg",
".",
"location",
")",
"else",
":",
"return",
"unicode",
"(",
"arg",
")"
] | if arg is an xblock . | train | false |
18,595 | def select_app_name():
name = utils.generate_app_name()
while App.objects.filter(id=name).exists():
name = utils.generate_app_name()
return name
| [
"def",
"select_app_name",
"(",
")",
":",
"name",
"=",
"utils",
".",
"generate_app_name",
"(",
")",
"while",
"App",
".",
"objects",
".",
"filter",
"(",
"id",
"=",
"name",
")",
".",
"exists",
"(",
")",
":",
"name",
"=",
"utils",
".",
"generate_app_name",
"(",
")",
"return",
"name"
] | select a unique randomly generated app name . | train | false |
18,596 | def log_conditional_probability(segmented_topics, per_topic_postings, num_docs):
m_lc = []
for s_i in segmented_topics:
for (w_prime, w_star) in s_i:
w_prime_docs = per_topic_postings[w_prime]
w_star_docs = per_topic_postings[w_star]
co_docs = w_prime_docs.intersection(w_star_docs)
if w_star_docs:
m_lc_i = np.log((((len(co_docs) / float(num_docs)) + EPSILON) / (len(w_star_docs) / float(num_docs))))
else:
m_lc_i = 0.0
m_lc.append(m_lc_i)
return m_lc
| [
"def",
"log_conditional_probability",
"(",
"segmented_topics",
",",
"per_topic_postings",
",",
"num_docs",
")",
":",
"m_lc",
"=",
"[",
"]",
"for",
"s_i",
"in",
"segmented_topics",
":",
"for",
"(",
"w_prime",
",",
"w_star",
")",
"in",
"s_i",
":",
"w_prime_docs",
"=",
"per_topic_postings",
"[",
"w_prime",
"]",
"w_star_docs",
"=",
"per_topic_postings",
"[",
"w_star",
"]",
"co_docs",
"=",
"w_prime_docs",
".",
"intersection",
"(",
"w_star_docs",
")",
"if",
"w_star_docs",
":",
"m_lc_i",
"=",
"np",
".",
"log",
"(",
"(",
"(",
"(",
"len",
"(",
"co_docs",
")",
"/",
"float",
"(",
"num_docs",
")",
")",
"+",
"EPSILON",
")",
"/",
"(",
"len",
"(",
"w_star_docs",
")",
"/",
"float",
"(",
"num_docs",
")",
")",
")",
")",
"else",
":",
"m_lc_i",
"=",
"0.0",
"m_lc",
".",
"append",
"(",
"m_lc_i",
")",
"return",
"m_lc"
] | this function calculates the log-conditional-probability measure which is used by coherence measures such as u_mass . | train | false |
18,597 | def TR0(rv):
return rv.normal().factor().expand()
| [
"def",
"TR0",
"(",
"rv",
")",
":",
"return",
"rv",
".",
"normal",
"(",
")",
".",
"factor",
"(",
")",
".",
"expand",
"(",
")"
] | simplification of rational polynomials . | train | false |
18,598 | def func_accepts_var_args(func):
return any((p for p in inspect.signature(func).parameters.values() if (p.kind == p.VAR_POSITIONAL)))
| [
"def",
"func_accepts_var_args",
"(",
"func",
")",
":",
"return",
"any",
"(",
"(",
"p",
"for",
"p",
"in",
"inspect",
".",
"signature",
"(",
"func",
")",
".",
"parameters",
".",
"values",
"(",
")",
"if",
"(",
"p",
".",
"kind",
"==",
"p",
".",
"VAR_POSITIONAL",
")",
")",
")"
] | return true if function func accepts positional arguments *args . | train | false |
18,599 | @click.group(cls=FlaskGroup, create_app=make_app, add_version_option=False)
@click.option('--config', expose_value=False, callback=set_config, required=False, is_flag=False, is_eager=True, metavar='CONFIG', help='Specify the config to use in dotted module notation e.g. flaskbb.configs.default.DefaultConfig')
@click.option('--version', expose_value=False, callback=get_version, is_flag=True, is_eager=True, help='Show the FlaskBB version.')
def flaskbb():
pass
| [
"@",
"click",
".",
"group",
"(",
"cls",
"=",
"FlaskGroup",
",",
"create_app",
"=",
"make_app",
",",
"add_version_option",
"=",
"False",
")",
"@",
"click",
".",
"option",
"(",
"'--config'",
",",
"expose_value",
"=",
"False",
",",
"callback",
"=",
"set_config",
",",
"required",
"=",
"False",
",",
"is_flag",
"=",
"False",
",",
"is_eager",
"=",
"True",
",",
"metavar",
"=",
"'CONFIG'",
",",
"help",
"=",
"'Specify the config to use in dotted module notation e.g. flaskbb.configs.default.DefaultConfig'",
")",
"@",
"click",
".",
"option",
"(",
"'--version'",
",",
"expose_value",
"=",
"False",
",",
"callback",
"=",
"get_version",
",",
"is_flag",
"=",
"True",
",",
"is_eager",
"=",
"True",
",",
"help",
"=",
"'Show the FlaskBB version.'",
")",
"def",
"flaskbb",
"(",
")",
":",
"pass"
] | this is the commandline interface for flaskbb . | train | false |
18,600 | def uniform_sequence(n):
return [random.uniform(0, n) for i in range(n)]
| [
"def",
"uniform_sequence",
"(",
"n",
")",
":",
"return",
"[",
"random",
".",
"uniform",
"(",
"0",
",",
"n",
")",
"for",
"i",
"in",
"range",
"(",
"n",
")",
"]"
] | return sample sequence of length n from a uniform distribution . | train | false |
18,601 | def test_extras_after_wheel(script, data):
simple = (script.site_packages / 'simple')
no_extra = script.pip('install', '--no-index', '-f', data.find_links, 'requires_simple_extra', expect_stderr=True)
assert (simple not in no_extra.files_created), no_extra.files_created
extra = script.pip('install', '--no-index', '-f', data.find_links, 'requires_simple_extra[extra]', expect_stderr=True)
assert (simple in extra.files_created), extra.files_created
| [
"def",
"test_extras_after_wheel",
"(",
"script",
",",
"data",
")",
":",
"simple",
"=",
"(",
"script",
".",
"site_packages",
"/",
"'simple'",
")",
"no_extra",
"=",
"script",
".",
"pip",
"(",
"'install'",
",",
"'--no-index'",
",",
"'-f'",
",",
"data",
".",
"find_links",
",",
"'requires_simple_extra'",
",",
"expect_stderr",
"=",
"True",
")",
"assert",
"(",
"simple",
"not",
"in",
"no_extra",
".",
"files_created",
")",
",",
"no_extra",
".",
"files_created",
"extra",
"=",
"script",
".",
"pip",
"(",
"'install'",
",",
"'--no-index'",
",",
"'-f'",
",",
"data",
".",
"find_links",
",",
"'requires_simple_extra[extra]'",
",",
"expect_stderr",
"=",
"True",
")",
"assert",
"(",
"simple",
"in",
"extra",
".",
"files_created",
")",
",",
"extra",
".",
"files_created"
] | test installing a package with extras after installing from a wheel . | train | false |
18,602 | def _base_ordering(base, degree):
base_len = len(base)
ordering = ([0] * degree)
for i in range(base_len):
ordering[base[i]] = i
current = base_len
for i in range(degree):
if (i not in base):
ordering[i] = current
current += 1
return ordering
| [
"def",
"_base_ordering",
"(",
"base",
",",
"degree",
")",
":",
"base_len",
"=",
"len",
"(",
"base",
")",
"ordering",
"=",
"(",
"[",
"0",
"]",
"*",
"degree",
")",
"for",
"i",
"in",
"range",
"(",
"base_len",
")",
":",
"ordering",
"[",
"base",
"[",
"i",
"]",
"]",
"=",
"i",
"current",
"=",
"base_len",
"for",
"i",
"in",
"range",
"(",
"degree",
")",
":",
"if",
"(",
"i",
"not",
"in",
"base",
")",
":",
"ordering",
"[",
"i",
"]",
"=",
"current",
"current",
"+=",
"1",
"return",
"ordering"
] | order {0 . | train | false |
18,603 | def get_active_users():
return frappe.db.sql(u"select count(*) from `tabUser`\n DCTB DCTB where enabled = 1 and user_type != 'Website User'\n DCTB DCTB and name not in ({})\n DCTB DCTB and hour(timediff(now(), last_active)) < 72".format(u', '.join(([u'%s'] * len(STANDARD_USERS)))), STANDARD_USERS)[0][0]
| [
"def",
"get_active_users",
"(",
")",
":",
"return",
"frappe",
".",
"db",
".",
"sql",
"(",
"u\"select count(*) from `tabUser`\\n DCTB DCTB where enabled = 1 and user_type != 'Website User'\\n DCTB DCTB and name not in ({})\\n DCTB DCTB and hour(timediff(now(), last_active)) < 72\"",
".",
"format",
"(",
"u', '",
".",
"join",
"(",
"(",
"[",
"u'%s'",
"]",
"*",
"len",
"(",
"STANDARD_USERS",
")",
")",
")",
")",
",",
"STANDARD_USERS",
")",
"[",
"0",
"]",
"[",
"0",
"]"
] | returns no . | train | false |
18,604 | @open_file(1, mode='wb')
def write_edgelist(G, path, comments='#', delimiter=' ', data=True, encoding='utf-8'):
for line in generate_edgelist(G, delimiter, data):
line += '\n'
path.write(line.encode(encoding))
| [
"@",
"open_file",
"(",
"1",
",",
"mode",
"=",
"'wb'",
")",
"def",
"write_edgelist",
"(",
"G",
",",
"path",
",",
"comments",
"=",
"'#'",
",",
"delimiter",
"=",
"' '",
",",
"data",
"=",
"True",
",",
"encoding",
"=",
"'utf-8'",
")",
":",
"for",
"line",
"in",
"generate_edgelist",
"(",
"G",
",",
"delimiter",
",",
"data",
")",
":",
"line",
"+=",
"'\\n'",
"path",
".",
"write",
"(",
"line",
".",
"encode",
"(",
"encoding",
")",
")"
] | write graph as a list of edges . | train | false |
18,605 | def enum(name, items, start=1, is_int=False):
enums = [(v, i) for (i, v) in enumerate(items, start)]
base = (pyenum.IntEnum if is_int else pyenum.Enum)
base = pyenum.unique(base)
return base(name, enums)
| [
"def",
"enum",
"(",
"name",
",",
"items",
",",
"start",
"=",
"1",
",",
"is_int",
"=",
"False",
")",
":",
"enums",
"=",
"[",
"(",
"v",
",",
"i",
")",
"for",
"(",
"i",
",",
"v",
")",
"in",
"enumerate",
"(",
"items",
",",
"start",
")",
"]",
"base",
"=",
"(",
"pyenum",
".",
"IntEnum",
"if",
"is_int",
"else",
"pyenum",
".",
"Enum",
")",
"base",
"=",
"pyenum",
".",
"unique",
"(",
"base",
")",
"return",
"base",
"(",
"name",
",",
"enums",
")"
] | construct a new enum object . | train | false |
18,606 | def get_bound(pts):
(x0, y0, x1, y1) = (INF, INF, (- INF), (- INF))
for (x, y) in pts:
x0 = min(x0, x)
y0 = min(y0, y)
x1 = max(x1, x)
y1 = max(y1, y)
return (x0, y0, x1, y1)
| [
"def",
"get_bound",
"(",
"pts",
")",
":",
"(",
"x0",
",",
"y0",
",",
"x1",
",",
"y1",
")",
"=",
"(",
"INF",
",",
"INF",
",",
"(",
"-",
"INF",
")",
",",
"(",
"-",
"INF",
")",
")",
"for",
"(",
"x",
",",
"y",
")",
"in",
"pts",
":",
"x0",
"=",
"min",
"(",
"x0",
",",
"x",
")",
"y0",
"=",
"min",
"(",
"y0",
",",
"y",
")",
"x1",
"=",
"max",
"(",
"x1",
",",
"x",
")",
"y1",
"=",
"max",
"(",
"y1",
",",
"y",
")",
"return",
"(",
"x0",
",",
"y0",
",",
"x1",
",",
"y1",
")"
] | compute a minimal rectangle that covers all the points . | train | true |
18,607 | @pytest.hookimpl(hookwrapper=True)
def pytest_sessionfinish(exitstatus):
outcome = (yield)
outcome.get_result()
cache_dir = os.path.join(os.path.abspath(os.path.dirname(__file__)), '..', '.cache')
try:
os.mkdir(cache_dir)
except FileExistsError:
pass
status_file = os.path.join(cache_dir, 'pytest_status')
with open(status_file, 'w', encoding='ascii') as f:
f.write(str(exitstatus))
| [
"@",
"pytest",
".",
"hookimpl",
"(",
"hookwrapper",
"=",
"True",
")",
"def",
"pytest_sessionfinish",
"(",
"exitstatus",
")",
":",
"outcome",
"=",
"(",
"yield",
")",
"outcome",
".",
"get_result",
"(",
")",
"cache_dir",
"=",
"os",
".",
"path",
".",
"join",
"(",
"os",
".",
"path",
".",
"abspath",
"(",
"os",
".",
"path",
".",
"dirname",
"(",
"__file__",
")",
")",
",",
"'..'",
",",
"'.cache'",
")",
"try",
":",
"os",
".",
"mkdir",
"(",
"cache_dir",
")",
"except",
"FileExistsError",
":",
"pass",
"status_file",
"=",
"os",
".",
"path",
".",
"join",
"(",
"cache_dir",
",",
"'pytest_status'",
")",
"with",
"open",
"(",
"status_file",
",",
"'w'",
",",
"encoding",
"=",
"'ascii'",
")",
"as",
"f",
":",
"f",
".",
"write",
"(",
"str",
"(",
"exitstatus",
")",
")"
] | create a file to tell run_pytest . | train | false |
18,609 | def simplelist(inline):
return paramfinder.findall(inline)
| [
"def",
"simplelist",
"(",
"inline",
")",
":",
"return",
"paramfinder",
".",
"findall",
"(",
"inline",
")"
] | parse a string to a list . | train | false |
18,611 | def getNewRepository():
return ExportRepository()
| [
"def",
"getNewRepository",
"(",
")",
":",
"return",
"ExportRepository",
"(",
")"
] | get new repository . | train | false |
18,612 | @default_selem
@pad_for_eccentric_selems
def opening(image, selem=None, out=None):
eroded = erosion(image, selem)
out = dilation(eroded, selem, out=out, shift_x=True, shift_y=True)
return out
| [
"@",
"default_selem",
"@",
"pad_for_eccentric_selems",
"def",
"opening",
"(",
"image",
",",
"selem",
"=",
"None",
",",
"out",
"=",
"None",
")",
":",
"eroded",
"=",
"erosion",
"(",
"image",
",",
"selem",
")",
"out",
"=",
"dilation",
"(",
"eroded",
",",
"selem",
",",
"out",
"=",
"out",
",",
"shift_x",
"=",
"True",
",",
"shift_y",
"=",
"True",
")",
"return",
"out"
] | return greyscale morphological opening of an image . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.