id_within_dataset int64 1 55.5k | snippet stringlengths 19 14.2k | tokens listlengths 6 1.63k | nl stringlengths 6 352 | split_within_dataset stringclasses 1 value | is_duplicated bool 2 classes |
|---|---|---|---|---|---|
35,658 | def list_vdirs(site, app=_DEFAULT_APP):
ret = dict()
pscmd = list()
pscmd.append("Get-WebVirtualDirectory -Site '{0}' -Application '{1}'".format(site, app))
pscmd.append(" | Select-Object PhysicalPath, @{ Name = 'name';")
pscmd.append(" Expression = { $_.path.Split('/')[-1] } }")
cmd_ret = _srvmgr(func=str().join(pscmd), as_json=True)
try:
items = json.loads(cmd_ret['stdout'], strict=False)
except ValueError:
_LOG.error('Unable to parse return data as Json.')
for item in items:
ret[item['name']] = {'sourcepath': item['physicalPath']}
if (not ret):
_LOG.warning('No vdirs found in output: %s', cmd_ret)
return ret
| [
"def",
"list_vdirs",
"(",
"site",
",",
"app",
"=",
"_DEFAULT_APP",
")",
":",
"ret",
"=",
"dict",
"(",
")",
"pscmd",
"=",
"list",
"(",
")",
"pscmd",
".",
"append",
"(",
"\"Get-WebVirtualDirectory -Site '{0}' -Application '{1}'\"",
".",
"format",
"(",
"site",
",",
"app",
")",
")",
"pscmd",
".",
"append",
"(",
"\" | Select-Object PhysicalPath, @{ Name = 'name';\"",
")",
"pscmd",
".",
"append",
"(",
"\" Expression = { $_.path.Split('/')[-1] } }\"",
")",
"cmd_ret",
"=",
"_srvmgr",
"(",
"func",
"=",
"str",
"(",
")",
".",
"join",
"(",
"pscmd",
")",
",",
"as_json",
"=",
"True",
")",
"try",
":",
"items",
"=",
"json",
".",
"loads",
"(",
"cmd_ret",
"[",
"'stdout'",
"]",
",",
"strict",
"=",
"False",
")",
"except",
"ValueError",
":",
"_LOG",
".",
"error",
"(",
"'Unable to parse return data as Json.'",
")",
"for",
"item",
"in",
"items",
":",
"ret",
"[",
"item",
"[",
"'name'",
"]",
"]",
"=",
"{",
"'sourcepath'",
":",
"item",
"[",
"'physicalPath'",
"]",
"}",
"if",
"(",
"not",
"ret",
")",
":",
"_LOG",
".",
"warning",
"(",
"'No vdirs found in output: %s'",
",",
"cmd_ret",
")",
"return",
"ret"
] | get all configured iis virtual directories for the specified site . | train | false |
35,659 | def expand_format_text(hosts, text):
return direct_format_text(expand_line(text, hosts))
| [
"def",
"expand_format_text",
"(",
"hosts",
",",
"text",
")",
":",
"return",
"direct_format_text",
"(",
"expand_line",
"(",
"text",
",",
"hosts",
")",
")"
] | following redirects in links . | train | false |
35,660 | def _MakeQuery(query, filters, orders):
clone = datastore_pb.Query()
clone.CopyFrom(query)
clone.clear_filter()
clone.clear_order()
clone.filter_list().extend(filters)
clone.order_list().extend(orders)
return datastore_query.Query._from_pb(clone)
| [
"def",
"_MakeQuery",
"(",
"query",
",",
"filters",
",",
"orders",
")",
":",
"clone",
"=",
"datastore_pb",
".",
"Query",
"(",
")",
"clone",
".",
"CopyFrom",
"(",
"query",
")",
"clone",
".",
"clear_filter",
"(",
")",
"clone",
".",
"clear_order",
"(",
")",
"clone",
".",
"filter_list",
"(",
")",
".",
"extend",
"(",
"filters",
")",
"clone",
".",
"order_list",
"(",
")",
".",
"extend",
"(",
"orders",
")",
"return",
"datastore_query",
".",
"Query",
".",
"_from_pb",
"(",
"clone",
")"
] | make a datastore_query . | train | false |
35,661 | def get_event_conditions(doctype, filters=None):
from frappe.desk.reportview import build_match_conditions
if (not frappe.has_permission(doctype)):
frappe.throw(_(u'Not Permitted'), frappe.PermissionError)
conditions = build_match_conditions(doctype)
conditions = ((conditions and (u' and ' + conditions)) or u'')
if filters:
filters = json.loads(filters)
for key in filters:
if filters[key]:
conditions += u'and `{0}` = "{1}"'.format(frappe.db.escape(key), frappe.db.escape(filters[key]))
return conditions
| [
"def",
"get_event_conditions",
"(",
"doctype",
",",
"filters",
"=",
"None",
")",
":",
"from",
"frappe",
".",
"desk",
".",
"reportview",
"import",
"build_match_conditions",
"if",
"(",
"not",
"frappe",
".",
"has_permission",
"(",
"doctype",
")",
")",
":",
"frappe",
".",
"throw",
"(",
"_",
"(",
"u'Not Permitted'",
")",
",",
"frappe",
".",
"PermissionError",
")",
"conditions",
"=",
"build_match_conditions",
"(",
"doctype",
")",
"conditions",
"=",
"(",
"(",
"conditions",
"and",
"(",
"u' and '",
"+",
"conditions",
")",
")",
"or",
"u''",
")",
"if",
"filters",
":",
"filters",
"=",
"json",
".",
"loads",
"(",
"filters",
")",
"for",
"key",
"in",
"filters",
":",
"if",
"filters",
"[",
"key",
"]",
":",
"conditions",
"+=",
"u'and `{0}` = \"{1}\"'",
".",
"format",
"(",
"frappe",
".",
"db",
".",
"escape",
"(",
"key",
")",
",",
"frappe",
".",
"db",
".",
"escape",
"(",
"filters",
"[",
"key",
"]",
")",
")",
"return",
"conditions"
] | returns sql conditions with user permissions and filters for event queries . | train | false |
35,665 | @mobile_template('questions/{mobile/}product_list.html')
def product_list(request, template):
return render(request, template, {'products': Product.objects.filter(questions_locales__locale=request.LANGUAGE_CODE)})
| [
"@",
"mobile_template",
"(",
"'questions/{mobile/}product_list.html'",
")",
"def",
"product_list",
"(",
"request",
",",
"template",
")",
":",
"return",
"render",
"(",
"request",
",",
"template",
",",
"{",
"'products'",
":",
"Product",
".",
"objects",
".",
"filter",
"(",
"questions_locales__locale",
"=",
"request",
".",
"LANGUAGE_CODE",
")",
"}",
")"
] | view to select a product to see related questions . | train | false |
35,666 | def _parse_record_data(entry_data):
ret = {}
ipv4addrs = []
aliases = []
if ('canonical' in entry_data):
ret['Canonical Name'] = entry_data['canonical']
if ('ipv4addrs' in entry_data):
for ipaddrs in entry_data['ipv4addrs']:
ipv4addrs.append(ipaddrs['ipv4addr'])
ret['IP Addresses'] = ipv4addrs
if ('ipv4addr' in entry_data):
ret['IP Address'] = entry_data['ipv4addr']
if ('aliases' in entry_data):
for alias in entry_data['aliases']:
aliases.append(alias)
ret['Aliases'] = aliases
if ('name' in entry_data):
ret['Name'] = entry_data['name']
if ('view' in entry_data):
ret['DNS View'] = entry_data['view']
if ('network_view' in entry_data):
ret['Network View'] = entry_data['network_view']
if ('comment' in entry_data):
ret['Comment'] = entry_data['comment']
if ('network' in entry_data):
ret['Network'] = entry_data['network']
if ('_ref' in entry_data):
ret['Record ID'] = entry_data['_ref']
return ret
| [
"def",
"_parse_record_data",
"(",
"entry_data",
")",
":",
"ret",
"=",
"{",
"}",
"ipv4addrs",
"=",
"[",
"]",
"aliases",
"=",
"[",
"]",
"if",
"(",
"'canonical'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'Canonical Name'",
"]",
"=",
"entry_data",
"[",
"'canonical'",
"]",
"if",
"(",
"'ipv4addrs'",
"in",
"entry_data",
")",
":",
"for",
"ipaddrs",
"in",
"entry_data",
"[",
"'ipv4addrs'",
"]",
":",
"ipv4addrs",
".",
"append",
"(",
"ipaddrs",
"[",
"'ipv4addr'",
"]",
")",
"ret",
"[",
"'IP Addresses'",
"]",
"=",
"ipv4addrs",
"if",
"(",
"'ipv4addr'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'IP Address'",
"]",
"=",
"entry_data",
"[",
"'ipv4addr'",
"]",
"if",
"(",
"'aliases'",
"in",
"entry_data",
")",
":",
"for",
"alias",
"in",
"entry_data",
"[",
"'aliases'",
"]",
":",
"aliases",
".",
"append",
"(",
"alias",
")",
"ret",
"[",
"'Aliases'",
"]",
"=",
"aliases",
"if",
"(",
"'name'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'Name'",
"]",
"=",
"entry_data",
"[",
"'name'",
"]",
"if",
"(",
"'view'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'DNS View'",
"]",
"=",
"entry_data",
"[",
"'view'",
"]",
"if",
"(",
"'network_view'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'Network View'",
"]",
"=",
"entry_data",
"[",
"'network_view'",
"]",
"if",
"(",
"'comment'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'Comment'",
"]",
"=",
"entry_data",
"[",
"'comment'",
"]",
"if",
"(",
"'network'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'Network'",
"]",
"=",
"entry_data",
"[",
"'network'",
"]",
"if",
"(",
"'_ref'",
"in",
"entry_data",
")",
":",
"ret",
"[",
"'Record ID'",
"]",
"=",
"entry_data",
"[",
"'_ref'",
"]",
"return",
"ret"
] | returns the right value data wed be interested in for the specified record type . | train | false |
35,667 | def InitDB(schema=None, callback=None, verify_or_create=True):
assert (not hasattr(DBClient, '_instance')), 'instance already initialized'
assert (schema is not None)
if options.options.localdb:
from local_client import LocalClient
DBClient.SetInstance(LocalClient(schema, read_only=options.options.readonly_db))
else:
from dynamodb_client import DynamoDBClient
DBClient._instance = DynamoDBClient(schema, read_only=options.options.readonly_db)
if verify_or_create:
schema.VerifyOrCreate(DBClient.Instance(), callback)
else:
callback([])
| [
"def",
"InitDB",
"(",
"schema",
"=",
"None",
",",
"callback",
"=",
"None",
",",
"verify_or_create",
"=",
"True",
")",
":",
"assert",
"(",
"not",
"hasattr",
"(",
"DBClient",
",",
"'_instance'",
")",
")",
",",
"'instance already initialized'",
"assert",
"(",
"schema",
"is",
"not",
"None",
")",
"if",
"options",
".",
"options",
".",
"localdb",
":",
"from",
"local_client",
"import",
"LocalClient",
"DBClient",
".",
"SetInstance",
"(",
"LocalClient",
"(",
"schema",
",",
"read_only",
"=",
"options",
".",
"options",
".",
"readonly_db",
")",
")",
"else",
":",
"from",
"dynamodb_client",
"import",
"DynamoDBClient",
"DBClient",
".",
"_instance",
"=",
"DynamoDBClient",
"(",
"schema",
",",
"read_only",
"=",
"options",
".",
"options",
".",
"readonly_db",
")",
"if",
"verify_or_create",
":",
"schema",
".",
"VerifyOrCreate",
"(",
"DBClient",
".",
"Instance",
"(",
")",
",",
"callback",
")",
"else",
":",
"callback",
"(",
"[",
"]",
")"
] | sets the db client instance . | train | false |
35,668 | def is_subclass(cls, classinfo):
try:
return issubclass(cls, classinfo)
except TypeError:
return False
| [
"def",
"is_subclass",
"(",
"cls",
",",
"classinfo",
")",
":",
"try",
":",
"return",
"issubclass",
"(",
"cls",
",",
"classinfo",
")",
"except",
"TypeError",
":",
"return",
"False"
] | a more sensible version of the issubclass builtin . | train | false |
35,669 | def extract_barcodes_from_mapping(labels):
barcodes = {}
re = compile('(\\w+) ([a-zA-Z0-9.]+) orig_bc=(\\w*) new_bc=\\w* bc_diffs=\\d+')
for label in labels:
tmatch = search(re, label)
flowgram_id = tmatch.group(2)
barcode = tmatch.group(3)
barcodes[flowgram_id] = barcode
return barcodes
| [
"def",
"extract_barcodes_from_mapping",
"(",
"labels",
")",
":",
"barcodes",
"=",
"{",
"}",
"re",
"=",
"compile",
"(",
"'(\\\\w+) ([a-zA-Z0-9.]+) orig_bc=(\\\\w*) new_bc=\\\\w* bc_diffs=\\\\d+'",
")",
"for",
"label",
"in",
"labels",
":",
"tmatch",
"=",
"search",
"(",
"re",
",",
"label",
")",
"flowgram_id",
"=",
"tmatch",
".",
"group",
"(",
"2",
")",
"barcode",
"=",
"tmatch",
".",
"group",
"(",
"3",
")",
"barcodes",
"[",
"flowgram_id",
"]",
"=",
"barcode",
"return",
"barcodes"
] | extract barcodes from split_libraries fasta headers . | train | false |
35,670 | def long_token():
hash = hashlib.sha1(shortuuid.uuid())
hash.update(settings.SECRET_KEY)
return hash.hexdigest()
| [
"def",
"long_token",
"(",
")",
":",
"hash",
"=",
"hashlib",
".",
"sha1",
"(",
"shortuuid",
".",
"uuid",
"(",
")",
")",
"hash",
".",
"update",
"(",
"settings",
".",
"SECRET_KEY",
")",
"return",
"hash",
".",
"hexdigest",
"(",
")"
] | generate a hash that can be used as an application secret . | train | true |
35,672 | def hash_password(password):
return hash_password_PBKDF2(password)
| [
"def",
"hash_password",
"(",
"password",
")",
":",
"return",
"hash_password_PBKDF2",
"(",
"password",
")"
] | hash a password . | train | false |
35,673 | def ready(zone):
ret = {'status': True}
res = __salt__['cmd.run_all']('zoneadm {zone} ready'.format(zone=('-u {0}'.format(zone) if _is_uuid(zone) else '-z {0}'.format(zone))))
ret['status'] = (res['retcode'] == 0)
ret['message'] = (res['stdout'] if ret['status'] else res['stderr'])
ret['message'] = ret['message'].replace('zoneadm: ', '')
if (ret['message'] == ''):
del ret['message']
return ret
| [
"def",
"ready",
"(",
"zone",
")",
":",
"ret",
"=",
"{",
"'status'",
":",
"True",
"}",
"res",
"=",
"__salt__",
"[",
"'cmd.run_all'",
"]",
"(",
"'zoneadm {zone} ready'",
".",
"format",
"(",
"zone",
"=",
"(",
"'-u {0}'",
".",
"format",
"(",
"zone",
")",
"if",
"_is_uuid",
"(",
"zone",
")",
"else",
"'-z {0}'",
".",
"format",
"(",
"zone",
")",
")",
")",
")",
"ret",
"[",
"'status'",
"]",
"=",
"(",
"res",
"[",
"'retcode'",
"]",
"==",
"0",
")",
"ret",
"[",
"'message'",
"]",
"=",
"(",
"res",
"[",
"'stdout'",
"]",
"if",
"ret",
"[",
"'status'",
"]",
"else",
"res",
"[",
"'stderr'",
"]",
")",
"ret",
"[",
"'message'",
"]",
"=",
"ret",
"[",
"'message'",
"]",
".",
"replace",
"(",
"'zoneadm: '",
",",
"''",
")",
"if",
"(",
"ret",
"[",
"'message'",
"]",
"==",
"''",
")",
":",
"del",
"ret",
"[",
"'message'",
"]",
"return",
"ret"
] | prepares a zone for running applications . | train | true |
35,674 | def equalContents(arr1, arr2):
return (frozenset(arr1) == frozenset(arr2))
| [
"def",
"equalContents",
"(",
"arr1",
",",
"arr2",
")",
":",
"return",
"(",
"frozenset",
"(",
"arr1",
")",
"==",
"frozenset",
"(",
"arr2",
")",
")"
] | checks if the set of unique elements of arr1 and arr2 are equivalent . | train | false |
35,676 | def generate_milestone_namespace(namespace, course_key=None):
if (namespace in NAMESPACE_CHOICES.values()):
if (namespace == 'entrance_exams'):
return '{}.{}'.format(unicode(course_key), NAMESPACE_CHOICES['ENTRANCE_EXAM'])
| [
"def",
"generate_milestone_namespace",
"(",
"namespace",
",",
"course_key",
"=",
"None",
")",
":",
"if",
"(",
"namespace",
"in",
"NAMESPACE_CHOICES",
".",
"values",
"(",
")",
")",
":",
"if",
"(",
"namespace",
"==",
"'entrance_exams'",
")",
":",
"return",
"'{}.{}'",
".",
"format",
"(",
"unicode",
"(",
"course_key",
")",
",",
"NAMESPACE_CHOICES",
"[",
"'ENTRANCE_EXAM'",
"]",
")"
] | returns a specifically-formatted namespace string for the specified type . | train | false |
35,677 | def call_on_class_only(*args, **kwargs):
raise AttributeError('This method should be called on a class object.')
| [
"def",
"call_on_class_only",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"raise",
"AttributeError",
"(",
"'This method should be called on a class object.'",
")"
] | raise exception when load methods are called on instance . | train | false |
35,678 | def cigar(individual):
return (((individual[0] ** 2) + (1000000.0 * sum(((gene * gene) for gene in individual)))),)
| [
"def",
"cigar",
"(",
"individual",
")",
":",
"return",
"(",
"(",
"(",
"individual",
"[",
"0",
"]",
"**",
"2",
")",
"+",
"(",
"1000000.0",
"*",
"sum",
"(",
"(",
"(",
"gene",
"*",
"gene",
")",
"for",
"gene",
"in",
"individual",
")",
")",
")",
")",
",",
")"
] | cigar test objective function . | train | false |
35,679 | def format_log(request, message_type, message):
from django_socketio.settings import MESSAGE_LOG_FORMAT
if (MESSAGE_LOG_FORMAT is None):
return None
now = datetime.now().replace(microsecond=0)
args = dict(request.META, TYPE=message_type, MESSAGE=message, TIME=now)
return ((MESSAGE_LOG_FORMAT % args) + '\n')
| [
"def",
"format_log",
"(",
"request",
",",
"message_type",
",",
"message",
")",
":",
"from",
"django_socketio",
".",
"settings",
"import",
"MESSAGE_LOG_FORMAT",
"if",
"(",
"MESSAGE_LOG_FORMAT",
"is",
"None",
")",
":",
"return",
"None",
"now",
"=",
"datetime",
".",
"now",
"(",
")",
".",
"replace",
"(",
"microsecond",
"=",
"0",
")",
"args",
"=",
"dict",
"(",
"request",
".",
"META",
",",
"TYPE",
"=",
"message_type",
",",
"MESSAGE",
"=",
"message",
",",
"TIME",
"=",
"now",
")",
"return",
"(",
"(",
"MESSAGE_LOG_FORMAT",
"%",
"args",
")",
"+",
"'\\n'",
")"
] | formats a log message similar to gevents pywsgi request logging . | train | true |
35,680 | def strFile(p, f, caseSensitive=True):
buf = type(p)()
buf_len = max(len(p), (2 ** (2 ** (2 ** 2))))
if (not caseSensitive):
p = p.lower()
while 1:
r = f.read((buf_len - len(p)))
if (not caseSensitive):
r = r.lower()
bytes_read = len(r)
if (bytes_read == 0):
return False
l = ((len(buf) + bytes_read) - buf_len)
if (l <= 0):
buf = (buf + r)
else:
buf = (buf[l:] + r)
if (buf.find(p) != (-1)):
return True
| [
"def",
"strFile",
"(",
"p",
",",
"f",
",",
"caseSensitive",
"=",
"True",
")",
":",
"buf",
"=",
"type",
"(",
"p",
")",
"(",
")",
"buf_len",
"=",
"max",
"(",
"len",
"(",
"p",
")",
",",
"(",
"2",
"**",
"(",
"2",
"**",
"(",
"2",
"**",
"2",
")",
")",
")",
")",
"if",
"(",
"not",
"caseSensitive",
")",
":",
"p",
"=",
"p",
".",
"lower",
"(",
")",
"while",
"1",
":",
"r",
"=",
"f",
".",
"read",
"(",
"(",
"buf_len",
"-",
"len",
"(",
"p",
")",
")",
")",
"if",
"(",
"not",
"caseSensitive",
")",
":",
"r",
"=",
"r",
".",
"lower",
"(",
")",
"bytes_read",
"=",
"len",
"(",
"r",
")",
"if",
"(",
"bytes_read",
"==",
"0",
")",
":",
"return",
"False",
"l",
"=",
"(",
"(",
"len",
"(",
"buf",
")",
"+",
"bytes_read",
")",
"-",
"buf_len",
")",
"if",
"(",
"l",
"<=",
"0",
")",
":",
"buf",
"=",
"(",
"buf",
"+",
"r",
")",
"else",
":",
"buf",
"=",
"(",
"buf",
"[",
"l",
":",
"]",
"+",
"r",
")",
"if",
"(",
"buf",
".",
"find",
"(",
"p",
")",
"!=",
"(",
"-",
"1",
")",
")",
":",
"return",
"True"
] | find whether string c{p} occurs in a read()able object c{f} . | train | false |
35,682 | def test_suggested_multiple_column_names(completer, complete_event):
text = u'SELECT id, from custom.products'
position = len(u'SELECT id, ')
result = set(completer.get_completions(Document(text=text, cursor_position=position), complete_event))
assert (set(result) == set(((testdata.columns(u'products', u'custom') + testdata.functions()) + list((testdata.builtin_functions() + testdata.keywords())))))
| [
"def",
"test_suggested_multiple_column_names",
"(",
"completer",
",",
"complete_event",
")",
":",
"text",
"=",
"u'SELECT id, from custom.products'",
"position",
"=",
"len",
"(",
"u'SELECT id, '",
")",
"result",
"=",
"set",
"(",
"completer",
".",
"get_completions",
"(",
"Document",
"(",
"text",
"=",
"text",
",",
"cursor_position",
"=",
"position",
")",
",",
"complete_event",
")",
")",
"assert",
"(",
"set",
"(",
"result",
")",
"==",
"set",
"(",
"(",
"(",
"testdata",
".",
"columns",
"(",
"u'products'",
",",
"u'custom'",
")",
"+",
"testdata",
".",
"functions",
"(",
")",
")",
"+",
"list",
"(",
"(",
"testdata",
".",
"builtin_functions",
"(",
")",
"+",
"testdata",
".",
"keywords",
"(",
")",
")",
")",
")",
")",
")"
] | suggest column and function names when selecting multiple columns from table . | train | false |
35,683 | def encryptARC(s):
iv = helpers.randomKey(8)
key = helpers.randomKey(8)
arc4main = ARC4.new(key)
encrypted = arc4main.encrypt(s)
return (encrypted, (key, iv))
| [
"def",
"encryptARC",
"(",
"s",
")",
":",
"iv",
"=",
"helpers",
".",
"randomKey",
"(",
"8",
")",
"key",
"=",
"helpers",
".",
"randomKey",
"(",
"8",
")",
"arc4main",
"=",
"ARC4",
".",
"new",
"(",
"key",
")",
"encrypted",
"=",
"arc4main",
".",
"encrypt",
"(",
"s",
")",
"return",
"(",
"encrypted",
",",
"(",
"key",
",",
"iv",
")",
")"
] | generates a random arc key and iv . | train | false |
35,684 | def fix_target(target, unit):
if (target == []):
return (target, [])
fixups = []
for (dummy, fix) in AUTOFIXES.items():
(target, fixed) = fix.fix_target(target, unit)
if fixed:
fixups.append(fix.name)
return (target, fixups)
| [
"def",
"fix_target",
"(",
"target",
",",
"unit",
")",
":",
"if",
"(",
"target",
"==",
"[",
"]",
")",
":",
"return",
"(",
"target",
",",
"[",
"]",
")",
"fixups",
"=",
"[",
"]",
"for",
"(",
"dummy",
",",
"fix",
")",
"in",
"AUTOFIXES",
".",
"items",
"(",
")",
":",
"(",
"target",
",",
"fixed",
")",
"=",
"fix",
".",
"fix_target",
"(",
"target",
",",
"unit",
")",
"if",
"fixed",
":",
"fixups",
".",
"append",
"(",
"fix",
".",
"name",
")",
"return",
"(",
"target",
",",
"fixups",
")"
] | apply each autofix to the target translation . | train | false |
35,685 | def _set_sleep_timer_service(service):
_apply_service(service, SonosDevice.set_sleep_timer, service.data[ATTR_SLEEP_TIME])
| [
"def",
"_set_sleep_timer_service",
"(",
"service",
")",
":",
"_apply_service",
"(",
"service",
",",
"SonosDevice",
".",
"set_sleep_timer",
",",
"service",
".",
"data",
"[",
"ATTR_SLEEP_TIME",
"]",
")"
] | set a timer . | train | false |
35,686 | def _FilesMatching(root, predicate=(lambda f: True)):
matches = []
for (path, _, files) in os.walk(root):
matches += [os.path.join(path, f) for f in files if predicate(f)]
return matches
| [
"def",
"_FilesMatching",
"(",
"root",
",",
"predicate",
"=",
"(",
"lambda",
"f",
":",
"True",
")",
")",
":",
"matches",
"=",
"[",
"]",
"for",
"(",
"path",
",",
"_",
",",
"files",
")",
"in",
"os",
".",
"walk",
"(",
"root",
")",
":",
"matches",
"+=",
"[",
"os",
".",
"path",
".",
"join",
"(",
"path",
",",
"f",
")",
"for",
"f",
"in",
"files",
"if",
"predicate",
"(",
"f",
")",
"]",
"return",
"matches"
] | finds all files under the given root that match the given predicate . | train | false |
35,687 | def fixAPI(document, url):
for node in domhelpers.findElementsWithAttribute(document, 'class', 'API'):
fullname = _getAPI(node)
anchor = dom.Element('a')
anchor.setAttribute('href', (url % (fullname,)))
anchor.setAttribute('title', fullname)
while node.childNodes:
child = node.childNodes[0]
node.removeChild(child)
anchor.appendChild(child)
node.appendChild(anchor)
if node.hasAttribute('base'):
node.removeAttribute('base')
| [
"def",
"fixAPI",
"(",
"document",
",",
"url",
")",
":",
"for",
"node",
"in",
"domhelpers",
".",
"findElementsWithAttribute",
"(",
"document",
",",
"'class'",
",",
"'API'",
")",
":",
"fullname",
"=",
"_getAPI",
"(",
"node",
")",
"anchor",
"=",
"dom",
".",
"Element",
"(",
"'a'",
")",
"anchor",
".",
"setAttribute",
"(",
"'href'",
",",
"(",
"url",
"%",
"(",
"fullname",
",",
")",
")",
")",
"anchor",
".",
"setAttribute",
"(",
"'title'",
",",
"fullname",
")",
"while",
"node",
".",
"childNodes",
":",
"child",
"=",
"node",
".",
"childNodes",
"[",
"0",
"]",
"node",
".",
"removeChild",
"(",
"child",
")",
"anchor",
".",
"appendChild",
"(",
"child",
")",
"node",
".",
"appendChild",
"(",
"anchor",
")",
"if",
"node",
".",
"hasAttribute",
"(",
"'base'",
")",
":",
"node",
".",
"removeAttribute",
"(",
"'base'",
")"
] | replace api references with links to api documentation . | train | false |
35,688 | def subscriptions_for(username, number=(-1), etag=None):
return gh.subscriptions_for(username, number, etag)
| [
"def",
"subscriptions_for",
"(",
"username",
",",
"number",
"=",
"(",
"-",
"1",
")",
",",
"etag",
"=",
"None",
")",
":",
"return",
"gh",
".",
"subscriptions_for",
"(",
"username",
",",
"number",
",",
"etag",
")"
] | iterate over repositories subscribed to by username . | train | false |
35,689 | def random_bytes_hex(bytes_length):
return hexlify(os.urandom(bytes_length)).decode('utf-8')
| [
"def",
"random_bytes_hex",
"(",
"bytes_length",
")",
":",
"return",
"hexlify",
"(",
"os",
".",
"urandom",
"(",
"bytes_length",
")",
")",
".",
"decode",
"(",
"'utf-8'",
")"
] | return a hexstring of bytes_length cryptographic-friendly random bytes . | train | false |
35,691 | def fill_in(at, basedate, units=adatetime.units):
if isinstance(at, datetime):
return at
args = {}
for unit in units:
v = getattr(at, unit)
if (v is None):
v = getattr(basedate, unit)
args[unit] = v
return fix(adatetime(**args))
| [
"def",
"fill_in",
"(",
"at",
",",
"basedate",
",",
"units",
"=",
"adatetime",
".",
"units",
")",
":",
"if",
"isinstance",
"(",
"at",
",",
"datetime",
")",
":",
"return",
"at",
"args",
"=",
"{",
"}",
"for",
"unit",
"in",
"units",
":",
"v",
"=",
"getattr",
"(",
"at",
",",
"unit",
")",
"if",
"(",
"v",
"is",
"None",
")",
":",
"v",
"=",
"getattr",
"(",
"basedate",
",",
"unit",
")",
"args",
"[",
"unit",
"]",
"=",
"v",
"return",
"fix",
"(",
"adatetime",
"(",
"**",
"args",
")",
")"
] | returns a copy of at with any unspecified units filled in with values from basedate . | train | false |
35,692 | def xml_dict(data):
xml = ''
ordered_items = OrderedDict(sorted(data.items()))
for (k, v) in ordered_items.items():
if isinstance(v, datetime.datetime):
v = date_to_str(v)
elif isinstance(v, (datetime.time, datetime.date)):
v = v.isoformat()
if (not isinstance(v, list)):
v = [v]
for value in v:
if isinstance(value, dict):
links = xml_add_links(value)
xml += ('<%s>' % k)
xml += xml_dict(value)
xml += links
xml += ('</%s>' % k)
else:
xml += ('<%s>%s</%s>' % (k, utils.escape(value), k))
return xml
| [
"def",
"xml_dict",
"(",
"data",
")",
":",
"xml",
"=",
"''",
"ordered_items",
"=",
"OrderedDict",
"(",
"sorted",
"(",
"data",
".",
"items",
"(",
")",
")",
")",
"for",
"(",
"k",
",",
"v",
")",
"in",
"ordered_items",
".",
"items",
"(",
")",
":",
"if",
"isinstance",
"(",
"v",
",",
"datetime",
".",
"datetime",
")",
":",
"v",
"=",
"date_to_str",
"(",
"v",
")",
"elif",
"isinstance",
"(",
"v",
",",
"(",
"datetime",
".",
"time",
",",
"datetime",
".",
"date",
")",
")",
":",
"v",
"=",
"v",
".",
"isoformat",
"(",
")",
"if",
"(",
"not",
"isinstance",
"(",
"v",
",",
"list",
")",
")",
":",
"v",
"=",
"[",
"v",
"]",
"for",
"value",
"in",
"v",
":",
"if",
"isinstance",
"(",
"value",
",",
"dict",
")",
":",
"links",
"=",
"xml_add_links",
"(",
"value",
")",
"xml",
"+=",
"(",
"'<%s>'",
"%",
"k",
")",
"xml",
"+=",
"xml_dict",
"(",
"value",
")",
"xml",
"+=",
"links",
"xml",
"+=",
"(",
"'</%s>'",
"%",
"k",
")",
"else",
":",
"xml",
"+=",
"(",
"'<%s>%s</%s>'",
"%",
"(",
"k",
",",
"utils",
".",
"escape",
"(",
"value",
")",
",",
"k",
")",
")",
"return",
"xml"
] | renders a dict as xml . | train | false |
35,693 | def unicode2utf8(arg):
if (PY2 and isinstance(arg, unicode)):
return arg.encode('utf-8')
return arg
| [
"def",
"unicode2utf8",
"(",
"arg",
")",
":",
"if",
"(",
"PY2",
"and",
"isinstance",
"(",
"arg",
",",
"unicode",
")",
")",
":",
"return",
"arg",
".",
"encode",
"(",
"'utf-8'",
")",
"return",
"arg"
] | only in python 2 . | train | false |
35,695 | def homepage(module=None, *match, **attr):
settings = current.deployment_settings
all_modules = settings.modules
layout = S3MainMenuDefaultLayout
c = ([module] + list(match))
if ('name' in attr):
name = attr['name']
attr.pop('name')
else:
if (module is None):
module = 'default'
if (module in all_modules):
m = all_modules[module]
name = m.name_nice
else:
name = module
if ('f' in attr):
f = attr['f']
del attr['f']
else:
f = 'index'
return layout(name, c=c, f=f, **attr)
| [
"def",
"homepage",
"(",
"module",
"=",
"None",
",",
"*",
"match",
",",
"**",
"attr",
")",
":",
"settings",
"=",
"current",
".",
"deployment_settings",
"all_modules",
"=",
"settings",
".",
"modules",
"layout",
"=",
"S3MainMenuDefaultLayout",
"c",
"=",
"(",
"[",
"module",
"]",
"+",
"list",
"(",
"match",
")",
")",
"if",
"(",
"'name'",
"in",
"attr",
")",
":",
"name",
"=",
"attr",
"[",
"'name'",
"]",
"attr",
".",
"pop",
"(",
"'name'",
")",
"else",
":",
"if",
"(",
"module",
"is",
"None",
")",
":",
"module",
"=",
"'default'",
"if",
"(",
"module",
"in",
"all_modules",
")",
":",
"m",
"=",
"all_modules",
"[",
"module",
"]",
"name",
"=",
"m",
".",
"name_nice",
"else",
":",
"name",
"=",
"module",
"if",
"(",
"'f'",
"in",
"attr",
")",
":",
"f",
"=",
"attr",
"[",
"'f'",
"]",
"del",
"attr",
"[",
"'f'",
"]",
"else",
":",
"f",
"=",
"'index'",
"return",
"layout",
"(",
"name",
",",
"c",
"=",
"c",
",",
"f",
"=",
"f",
",",
"**",
"attr",
")"
] | shortcut for module homepage menu items using the mm layout . | train | false |
35,696 | def getDBTables(uri=None):
DB_TABLES = []
for table in DB_SCHEMA:
if (table.name in TABLES_REPOSITORY):
DB_TABLES.append(TABLES_REPOSITORY[table.name])
continue
attrs = {'_imdbpyName': table.name, '_imdbpySchema': table, 'addIndexes': addIndexes, 'addForeignKeys': addForeignKeys}
for col in table.cols:
if (col.name == 'id'):
continue
attrs[col.name] = MAP_COLS[col.kind](**col.params)
cls = type(table.name, (SQLObject,), attrs)
DB_TABLES.append(cls)
TABLES_REPOSITORY[table.name] = cls
return DB_TABLES
| [
"def",
"getDBTables",
"(",
"uri",
"=",
"None",
")",
":",
"DB_TABLES",
"=",
"[",
"]",
"for",
"table",
"in",
"DB_SCHEMA",
":",
"if",
"(",
"table",
".",
"name",
"in",
"TABLES_REPOSITORY",
")",
":",
"DB_TABLES",
".",
"append",
"(",
"TABLES_REPOSITORY",
"[",
"table",
".",
"name",
"]",
")",
"continue",
"attrs",
"=",
"{",
"'_imdbpyName'",
":",
"table",
".",
"name",
",",
"'_imdbpySchema'",
":",
"table",
",",
"'addIndexes'",
":",
"addIndexes",
",",
"'addForeignKeys'",
":",
"addForeignKeys",
"}",
"for",
"col",
"in",
"table",
".",
"cols",
":",
"if",
"(",
"col",
".",
"name",
"==",
"'id'",
")",
":",
"continue",
"attrs",
"[",
"col",
".",
"name",
"]",
"=",
"MAP_COLS",
"[",
"col",
".",
"kind",
"]",
"(",
"**",
"col",
".",
"params",
")",
"cls",
"=",
"type",
"(",
"table",
".",
"name",
",",
"(",
"SQLObject",
",",
")",
",",
"attrs",
")",
"DB_TABLES",
".",
"append",
"(",
"cls",
")",
"TABLES_REPOSITORY",
"[",
"table",
".",
"name",
"]",
"=",
"cls",
"return",
"DB_TABLES"
] | return a list of classes to be used to access the database through the sqlobject orm . | train | false |
35,697 | def checkSum():
val = 0
for (root, dirs, files) in os.walk(os.getcwd()):
for extension in EXTENSIONS:
for f in fnmatch.filter(files, extension):
stats = os.stat(os.path.join(root, f))
val += (stats[stat.ST_SIZE] + stats[stat.ST_MTIME])
return val
| [
"def",
"checkSum",
"(",
")",
":",
"val",
"=",
"0",
"for",
"(",
"root",
",",
"dirs",
",",
"files",
")",
"in",
"os",
".",
"walk",
"(",
"os",
".",
"getcwd",
"(",
")",
")",
":",
"for",
"extension",
"in",
"EXTENSIONS",
":",
"for",
"f",
"in",
"fnmatch",
".",
"filter",
"(",
"files",
",",
"extension",
")",
":",
"stats",
"=",
"os",
".",
"stat",
"(",
"os",
".",
"path",
".",
"join",
"(",
"root",
",",
"f",
")",
")",
"val",
"+=",
"(",
"stats",
"[",
"stat",
".",
"ST_SIZE",
"]",
"+",
"stats",
"[",
"stat",
".",
"ST_MTIME",
"]",
")",
"return",
"val"
] | return a long which can be used to know if any . | train | false |
35,698 | def mock_tweet():
count = random.randint(70, 140)
return ''.join([random.choice(string.ascii_letters) for _ in range(count)])
| [
"def",
"mock_tweet",
"(",
")",
":",
"count",
"=",
"random",
".",
"randint",
"(",
"70",
",",
"140",
")",
"return",
"''",
".",
"join",
"(",
"[",
"random",
".",
"choice",
"(",
"string",
".",
"ascii_letters",
")",
"for",
"_",
"in",
"range",
"(",
"count",
")",
"]",
")"
] | generate some random tweet text . | train | false |
35,700 | def get_poetry(address):
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
sock.connect(address)
poem = ''
while True:
data = sock.recv(1024)
if (not data):
sock.close()
break
poem += data
return poem
| [
"def",
"get_poetry",
"(",
"address",
")",
":",
"sock",
"=",
"socket",
".",
"socket",
"(",
"socket",
".",
"AF_INET",
",",
"socket",
".",
"SOCK_STREAM",
")",
"sock",
".",
"connect",
"(",
"address",
")",
"poem",
"=",
"''",
"while",
"True",
":",
"data",
"=",
"sock",
".",
"recv",
"(",
"1024",
")",
"if",
"(",
"not",
"data",
")",
":",
"sock",
".",
"close",
"(",
")",
"break",
"poem",
"+=",
"data",
"return",
"poem"
] | download a piece of poetry from the given address . | train | false |
35,701 | def saml_metadata_view(request):
if (not SAMLConfiguration.is_enabled(request.site)):
raise Http404
complete_url = reverse('social:complete', args=('tpa-saml',))
if (settings.APPEND_SLASH and (not complete_url.endswith('/'))):
complete_url = (complete_url + '/')
saml_backend = load_backend(load_strategy(request), 'tpa-saml', redirect_uri=complete_url)
(metadata, errors) = saml_backend.generate_metadata_xml()
if (not errors):
return HttpResponse(content=metadata, content_type='text/xml')
return HttpResponseServerError(content=', '.join(errors))
| [
"def",
"saml_metadata_view",
"(",
"request",
")",
":",
"if",
"(",
"not",
"SAMLConfiguration",
".",
"is_enabled",
"(",
"request",
".",
"site",
")",
")",
":",
"raise",
"Http404",
"complete_url",
"=",
"reverse",
"(",
"'social:complete'",
",",
"args",
"=",
"(",
"'tpa-saml'",
",",
")",
")",
"if",
"(",
"settings",
".",
"APPEND_SLASH",
"and",
"(",
"not",
"complete_url",
".",
"endswith",
"(",
"'/'",
")",
")",
")",
":",
"complete_url",
"=",
"(",
"complete_url",
"+",
"'/'",
")",
"saml_backend",
"=",
"load_backend",
"(",
"load_strategy",
"(",
"request",
")",
",",
"'tpa-saml'",
",",
"redirect_uri",
"=",
"complete_url",
")",
"(",
"metadata",
",",
"errors",
")",
"=",
"saml_backend",
".",
"generate_metadata_xml",
"(",
")",
"if",
"(",
"not",
"errors",
")",
":",
"return",
"HttpResponse",
"(",
"content",
"=",
"metadata",
",",
"content_type",
"=",
"'text/xml'",
")",
"return",
"HttpResponseServerError",
"(",
"content",
"=",
"', '",
".",
"join",
"(",
"errors",
")",
")"
] | get the service provider metadata for this edx-platform instance . | train | false |
35,702 | def _buildNetwork(*layers, **options):
bias = (options['bias'] if ('bias' in options) else False)
net = FeedForwardNetwork()
layerParts = iter(layers)
firstPart = iter(next(layerParts))
firstLayer = next(firstPart)
net.addInputModule(firstLayer)
prevLayer = firstLayer
for part in chain(firstPart, layerParts):
new_part = True
for layer in part:
net.addModule(layer)
if new_part:
ConnectionClass = FullConnection
if bias:
biasUnit = BiasUnit(('BiasUnit for %s' % layer.name))
net.addModule(biasUnit)
net.addConnection(FullConnection(biasUnit, layer))
else:
ConnectionClass = IdentityConnection
new_part = False
conn = ConnectionClass(prevLayer, layer)
net.addConnection(conn)
prevLayer = layer
net.addOutputModule(layer)
net.sortModules()
return net
| [
"def",
"_buildNetwork",
"(",
"*",
"layers",
",",
"**",
"options",
")",
":",
"bias",
"=",
"(",
"options",
"[",
"'bias'",
"]",
"if",
"(",
"'bias'",
"in",
"options",
")",
"else",
"False",
")",
"net",
"=",
"FeedForwardNetwork",
"(",
")",
"layerParts",
"=",
"iter",
"(",
"layers",
")",
"firstPart",
"=",
"iter",
"(",
"next",
"(",
"layerParts",
")",
")",
"firstLayer",
"=",
"next",
"(",
"firstPart",
")",
"net",
".",
"addInputModule",
"(",
"firstLayer",
")",
"prevLayer",
"=",
"firstLayer",
"for",
"part",
"in",
"chain",
"(",
"firstPart",
",",
"layerParts",
")",
":",
"new_part",
"=",
"True",
"for",
"layer",
"in",
"part",
":",
"net",
".",
"addModule",
"(",
"layer",
")",
"if",
"new_part",
":",
"ConnectionClass",
"=",
"FullConnection",
"if",
"bias",
":",
"biasUnit",
"=",
"BiasUnit",
"(",
"(",
"'BiasUnit for %s'",
"%",
"layer",
".",
"name",
")",
")",
"net",
".",
"addModule",
"(",
"biasUnit",
")",
"net",
".",
"addConnection",
"(",
"FullConnection",
"(",
"biasUnit",
",",
"layer",
")",
")",
"else",
":",
"ConnectionClass",
"=",
"IdentityConnection",
"new_part",
"=",
"False",
"conn",
"=",
"ConnectionClass",
"(",
"prevLayer",
",",
"layer",
")",
"net",
".",
"addConnection",
"(",
"conn",
")",
"prevLayer",
"=",
"layer",
"net",
".",
"addOutputModule",
"(",
"layer",
")",
"net",
".",
"sortModules",
"(",
")",
"return",
"net"
] | this is a helper function to create different kinds of networks . | train | false |
35,703 | def get_infos(dir_name):
infos = [i for i in iter_docstring_info(dir_name)]
infos.sort(key=(lambda x: x['source']))
for (num, info) in enumerate(infos):
info['num'] = num
enhance_info_description(info)
return infos
| [
"def",
"get_infos",
"(",
"dir_name",
")",
":",
"infos",
"=",
"[",
"i",
"for",
"i",
"in",
"iter_docstring_info",
"(",
"dir_name",
")",
"]",
"infos",
".",
"sort",
"(",
"key",
"=",
"(",
"lambda",
"x",
":",
"x",
"[",
"'source'",
"]",
")",
")",
"for",
"(",
"num",
",",
"info",
")",
"in",
"enumerate",
"(",
"infos",
")",
":",
"info",
"[",
"'num'",
"]",
"=",
"num",
"enhance_info_description",
"(",
"info",
")",
"return",
"infos"
] | return infos . | train | false |
35,704 | def at(*args, **kwargs):
if (len(args) < 2):
return {'jobs': []}
if ('tag' in kwargs):
stdin = '### SALT: {0}\n{1}'.format(kwargs['tag'], ' '.join(args[1:]))
else:
stdin = ' '.join(args[1:])
cmd_kwargs = {'stdin': stdin, 'python_shell': False}
if ('runas' in kwargs):
cmd_kwargs['runas'] = kwargs['runas']
res = __salt__['cmd.run_all']('at "{timespec}"'.format(timespec=args[0]), **cmd_kwargs)
if (res['retcode'] > 0):
if ('bad time specification' in res['stderr']):
return {'jobs': [], 'error': 'invalid timespec'}
return {'jobs': [], 'error': res['stderr']}
else:
jobid = res['stderr'].splitlines()[1]
jobid = str(jobid.split()[1])
return atq(jobid)
| [
"def",
"at",
"(",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"if",
"(",
"len",
"(",
"args",
")",
"<",
"2",
")",
":",
"return",
"{",
"'jobs'",
":",
"[",
"]",
"}",
"if",
"(",
"'tag'",
"in",
"kwargs",
")",
":",
"stdin",
"=",
"'### SALT: {0}\\n{1}'",
".",
"format",
"(",
"kwargs",
"[",
"'tag'",
"]",
",",
"' '",
".",
"join",
"(",
"args",
"[",
"1",
":",
"]",
")",
")",
"else",
":",
"stdin",
"=",
"' '",
".",
"join",
"(",
"args",
"[",
"1",
":",
"]",
")",
"cmd_kwargs",
"=",
"{",
"'stdin'",
":",
"stdin",
",",
"'python_shell'",
":",
"False",
"}",
"if",
"(",
"'runas'",
"in",
"kwargs",
")",
":",
"cmd_kwargs",
"[",
"'runas'",
"]",
"=",
"kwargs",
"[",
"'runas'",
"]",
"res",
"=",
"__salt__",
"[",
"'cmd.run_all'",
"]",
"(",
"'at \"{timespec}\"'",
".",
"format",
"(",
"timespec",
"=",
"args",
"[",
"0",
"]",
")",
",",
"**",
"cmd_kwargs",
")",
"if",
"(",
"res",
"[",
"'retcode'",
"]",
">",
"0",
")",
":",
"if",
"(",
"'bad time specification'",
"in",
"res",
"[",
"'stderr'",
"]",
")",
":",
"return",
"{",
"'jobs'",
":",
"[",
"]",
",",
"'error'",
":",
"'invalid timespec'",
"}",
"return",
"{",
"'jobs'",
":",
"[",
"]",
",",
"'error'",
":",
"res",
"[",
"'stderr'",
"]",
"}",
"else",
":",
"jobid",
"=",
"res",
"[",
"'stderr'",
"]",
".",
"splitlines",
"(",
")",
"[",
"1",
"]",
"jobid",
"=",
"str",
"(",
"jobid",
".",
"split",
"(",
")",
"[",
"1",
"]",
")",
"return",
"atq",
"(",
"jobid",
")"
] | usage: {{ dictionary|at:key}} . | train | true |
35,705 | @pytest.mark.usefixtures('clean_system', 'remove_additional_dirs')
def test_cookiecutter_templated_context():
main.cookiecutter('tests/fake-repo-tmpl', no_input=True)
assert os.path.isdir('fake-project-templated')
| [
"@",
"pytest",
".",
"mark",
".",
"usefixtures",
"(",
"'clean_system'",
",",
"'remove_additional_dirs'",
")",
"def",
"test_cookiecutter_templated_context",
"(",
")",
":",
"main",
".",
"cookiecutter",
"(",
"'tests/fake-repo-tmpl'",
",",
"no_input",
"=",
"True",
")",
"assert",
"os",
".",
"path",
".",
"isdir",
"(",
"'fake-project-templated'",
")"
] | call cookiecutter() with no_input=true and templates in the cookiecutter . | train | false |
35,707 | def new(rsa_key):
return PKCS115_SigScheme(rsa_key)
| [
"def",
"new",
"(",
"rsa_key",
")",
":",
"return",
"PKCS115_SigScheme",
"(",
"rsa_key",
")"
] | return a form for a new genericimagemodeljob . | train | false |
35,708 | def load_pack_index(path):
with GitFile(path, 'rb') as f:
return load_pack_index_file(path, f)
| [
"def",
"load_pack_index",
"(",
"path",
")",
":",
"with",
"GitFile",
"(",
"path",
",",
"'rb'",
")",
"as",
"f",
":",
"return",
"load_pack_index_file",
"(",
"path",
",",
"f",
")"
] | load an index file by path . | train | false |
35,709 | def places_autocomplete_query(client, input_text, offset=None, location=None, radius=None, language=None):
return _autocomplete(client, 'query', input_text, offset=offset, location=location, radius=radius, language=language)
| [
"def",
"places_autocomplete_query",
"(",
"client",
",",
"input_text",
",",
"offset",
"=",
"None",
",",
"location",
"=",
"None",
",",
"radius",
"=",
"None",
",",
"language",
"=",
"None",
")",
":",
"return",
"_autocomplete",
"(",
"client",
",",
"'query'",
",",
"input_text",
",",
"offset",
"=",
"offset",
",",
"location",
"=",
"location",
",",
"radius",
"=",
"radius",
",",
"language",
"=",
"language",
")"
] | returns place predictions given a textual search query . | train | true |
35,710 | def create_text_record(text):
opos = text.tell()
text.seek(0, 2)
npos = min(((opos + RECORD_SIZE), text.tell()))
extra = 0
last = ''
while (not last.decode(u'utf-8', u'ignore')):
size = (len(last) + 1)
text.seek((npos - size))
last = text.read(size)
try:
last.decode(u'utf-8', u'strict')
except UnicodeDecodeError:
prev = len(last)
while True:
text.seek((npos - prev))
last = text.read((len(last) + 1))
try:
last.decode(u'utf-8')
except UnicodeDecodeError:
pass
else:
break
extra = (len(last) - prev)
text.seek(opos)
data = text.read(RECORD_SIZE)
overlap = text.read(extra)
text.seek(npos)
return (data, overlap)
| [
"def",
"create_text_record",
"(",
"text",
")",
":",
"opos",
"=",
"text",
".",
"tell",
"(",
")",
"text",
".",
"seek",
"(",
"0",
",",
"2",
")",
"npos",
"=",
"min",
"(",
"(",
"(",
"opos",
"+",
"RECORD_SIZE",
")",
",",
"text",
".",
"tell",
"(",
")",
")",
")",
"extra",
"=",
"0",
"last",
"=",
"''",
"while",
"(",
"not",
"last",
".",
"decode",
"(",
"u'utf-8'",
",",
"u'ignore'",
")",
")",
":",
"size",
"=",
"(",
"len",
"(",
"last",
")",
"+",
"1",
")",
"text",
".",
"seek",
"(",
"(",
"npos",
"-",
"size",
")",
")",
"last",
"=",
"text",
".",
"read",
"(",
"size",
")",
"try",
":",
"last",
".",
"decode",
"(",
"u'utf-8'",
",",
"u'strict'",
")",
"except",
"UnicodeDecodeError",
":",
"prev",
"=",
"len",
"(",
"last",
")",
"while",
"True",
":",
"text",
".",
"seek",
"(",
"(",
"npos",
"-",
"prev",
")",
")",
"last",
"=",
"text",
".",
"read",
"(",
"(",
"len",
"(",
"last",
")",
"+",
"1",
")",
")",
"try",
":",
"last",
".",
"decode",
"(",
"u'utf-8'",
")",
"except",
"UnicodeDecodeError",
":",
"pass",
"else",
":",
"break",
"extra",
"=",
"(",
"len",
"(",
"last",
")",
"-",
"prev",
")",
"text",
".",
"seek",
"(",
"opos",
")",
"data",
"=",
"text",
".",
"read",
"(",
"RECORD_SIZE",
")",
"overlap",
"=",
"text",
".",
"read",
"(",
"extra",
")",
"text",
".",
"seek",
"(",
"npos",
")",
"return",
"(",
"data",
",",
"overlap",
")"
] | return a palmdoc record of size record_size from the text file object . | train | false |
35,712 | def ymax(y):
y1 = y.ravel()
maxy1 = np.amax(y1)
e1 = np.exp((y1 - maxy1))
p1 = (e1 / np.sum(e1))
y1 = np.log((1e-20 + p1))
ix = np.argmax(y1)
return (ix, y1[ix])
| [
"def",
"ymax",
"(",
"y",
")",
":",
"y1",
"=",
"y",
".",
"ravel",
"(",
")",
"maxy1",
"=",
"np",
".",
"amax",
"(",
"y1",
")",
"e1",
"=",
"np",
".",
"exp",
"(",
"(",
"y1",
"-",
"maxy1",
")",
")",
"p1",
"=",
"(",
"e1",
"/",
"np",
".",
"sum",
"(",
"e1",
")",
")",
"y1",
"=",
"np",
".",
"log",
"(",
"(",
"1e-20",
"+",
"p1",
")",
")",
"ix",
"=",
"np",
".",
"argmax",
"(",
"y1",
")",
"return",
"(",
"ix",
",",
"y1",
"[",
"ix",
"]",
")"
] | simple helper function here that takes unnormalized logprobs . | train | false |
35,713 | def _GenDiscoveryDocCallback(args, discovery_func=GenDiscoveryDoc):
(services, doc_format, output_path, hostname) = (args.service, args.format, args.output, args.hostname)
discovery_paths = discovery_func(services, doc_format, output_path, hostname=hostname)
for discovery_path in discovery_paths:
print ('API discovery document written to %s' % discovery_path)
| [
"def",
"_GenDiscoveryDocCallback",
"(",
"args",
",",
"discovery_func",
"=",
"GenDiscoveryDoc",
")",
":",
"(",
"services",
",",
"doc_format",
",",
"output_path",
",",
"hostname",
")",
"=",
"(",
"args",
".",
"service",
",",
"args",
".",
"format",
",",
"args",
".",
"output",
",",
"args",
".",
"hostname",
")",
"discovery_paths",
"=",
"discovery_func",
"(",
"services",
",",
"doc_format",
",",
"output_path",
",",
"hostname",
"=",
"hostname",
")",
"for",
"discovery_path",
"in",
"discovery_paths",
":",
"print",
"(",
"'API discovery document written to %s'",
"%",
"discovery_path",
")"
] | generate discovery docs to files . | train | false |
35,715 | def unequal_vals_constraint(A, a, B, b):
return (a != b)
| [
"def",
"unequal_vals_constraint",
"(",
"A",
",",
"a",
",",
"B",
",",
"b",
")",
":",
"return",
"(",
"a",
"!=",
"b",
")"
] | a simple constraint: two neighbors must always have different values . | train | false |
35,716 | def from_dataset(dataset, num_examples):
if (dataset.view_converter is not None):
try:
(V, y) = dataset.get_batch_topo(num_examples, True)
except TypeError:
if (isinstance(dataset, DenseDesignMatrix) and (dataset.X is None) and (not control.get_load_data())):
warnings.warn("from_dataset wasn't able to make subset of dataset, using the whole thing")
return DenseDesignMatrix(X=None, view_converter=dataset.view_converter)
raise
rval = DenseDesignMatrix(topo_view=V, y=y, y_labels=dataset.y_labels)
rval.adjust_for_viewer = dataset.adjust_for_viewer
else:
(X, y) = dataset.get_batch_design(num_examples, True)
rval = DenseDesignMatrix(X=X, y=y, y_labels=dataset.y_labels)
return rval
| [
"def",
"from_dataset",
"(",
"dataset",
",",
"num_examples",
")",
":",
"if",
"(",
"dataset",
".",
"view_converter",
"is",
"not",
"None",
")",
":",
"try",
":",
"(",
"V",
",",
"y",
")",
"=",
"dataset",
".",
"get_batch_topo",
"(",
"num_examples",
",",
"True",
")",
"except",
"TypeError",
":",
"if",
"(",
"isinstance",
"(",
"dataset",
",",
"DenseDesignMatrix",
")",
"and",
"(",
"dataset",
".",
"X",
"is",
"None",
")",
"and",
"(",
"not",
"control",
".",
"get_load_data",
"(",
")",
")",
")",
":",
"warnings",
".",
"warn",
"(",
"\"from_dataset wasn't able to make subset of dataset, using the whole thing\"",
")",
"return",
"DenseDesignMatrix",
"(",
"X",
"=",
"None",
",",
"view_converter",
"=",
"dataset",
".",
"view_converter",
")",
"raise",
"rval",
"=",
"DenseDesignMatrix",
"(",
"topo_view",
"=",
"V",
",",
"y",
"=",
"y",
",",
"y_labels",
"=",
"dataset",
".",
"y_labels",
")",
"rval",
".",
"adjust_for_viewer",
"=",
"dataset",
".",
"adjust_for_viewer",
"else",
":",
"(",
"X",
",",
"y",
")",
"=",
"dataset",
".",
"get_batch_design",
"(",
"num_examples",
",",
"True",
")",
"rval",
"=",
"DenseDesignMatrix",
"(",
"X",
"=",
"X",
",",
"y",
"=",
"y",
",",
"y_labels",
"=",
"dataset",
".",
"y_labels",
")",
"return",
"rval"
] | constructs a random subset of a densedesignmatrix parameters dataset : densedesignmatrix num_examples : int returns sub_dataset : densedesignmatrix a new dataset containing num_examples examples . | train | false |
35,717 | def mk_test_name(name, value, index=0):
trivial_types = (type(None), bool, str, int, float)
try:
trivial_types += (unicode,)
except NameError:
pass
def is_trivial(value):
if isinstance(value, trivial_types):
return True
if isinstance(value, (list, tuple)):
return all(map(is_trivial, value))
return False
if (is_hash_randomized() and (not is_trivial(value))):
return '{0}_{1}'.format(name, (index + 1))
try:
value = str(value)
except UnicodeEncodeError:
value = value.encode('ascii', 'backslashreplace')
test_name = '{0}_{1}_{2}'.format(name, (index + 1), value)
return re.sub('\\W|^(?=\\d)', '_', test_name)
| [
"def",
"mk_test_name",
"(",
"name",
",",
"value",
",",
"index",
"=",
"0",
")",
":",
"trivial_types",
"=",
"(",
"type",
"(",
"None",
")",
",",
"bool",
",",
"str",
",",
"int",
",",
"float",
")",
"try",
":",
"trivial_types",
"+=",
"(",
"unicode",
",",
")",
"except",
"NameError",
":",
"pass",
"def",
"is_trivial",
"(",
"value",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"trivial_types",
")",
":",
"return",
"True",
"if",
"isinstance",
"(",
"value",
",",
"(",
"list",
",",
"tuple",
")",
")",
":",
"return",
"all",
"(",
"map",
"(",
"is_trivial",
",",
"value",
")",
")",
"return",
"False",
"if",
"(",
"is_hash_randomized",
"(",
")",
"and",
"(",
"not",
"is_trivial",
"(",
"value",
")",
")",
")",
":",
"return",
"'{0}_{1}'",
".",
"format",
"(",
"name",
",",
"(",
"index",
"+",
"1",
")",
")",
"try",
":",
"value",
"=",
"str",
"(",
"value",
")",
"except",
"UnicodeEncodeError",
":",
"value",
"=",
"value",
".",
"encode",
"(",
"'ascii'",
",",
"'backslashreplace'",
")",
"test_name",
"=",
"'{0}_{1}_{2}'",
".",
"format",
"(",
"name",
",",
"(",
"index",
"+",
"1",
")",
",",
"value",
")",
"return",
"re",
".",
"sub",
"(",
"'\\\\W|^(?=\\\\d)'",
",",
"'_'",
",",
"test_name",
")"
] | generate a new name for a test case . | train | false |
35,718 | def apply_request_extensions(request, extensions=None):
if (extensions is None):
extensions = request.registry.queryUtility(IRequestExtensions)
if (extensions is not None):
for (name, fn) in iteritems_(extensions.methods):
method = fn.__get__(request, request.__class__)
setattr(request, name, method)
InstancePropertyHelper.apply_properties(request, extensions.descriptors)
| [
"def",
"apply_request_extensions",
"(",
"request",
",",
"extensions",
"=",
"None",
")",
":",
"if",
"(",
"extensions",
"is",
"None",
")",
":",
"extensions",
"=",
"request",
".",
"registry",
".",
"queryUtility",
"(",
"IRequestExtensions",
")",
"if",
"(",
"extensions",
"is",
"not",
"None",
")",
":",
"for",
"(",
"name",
",",
"fn",
")",
"in",
"iteritems_",
"(",
"extensions",
".",
"methods",
")",
":",
"method",
"=",
"fn",
".",
"__get__",
"(",
"request",
",",
"request",
".",
"__class__",
")",
"setattr",
"(",
"request",
",",
"name",
",",
"method",
")",
"InstancePropertyHelper",
".",
"apply_properties",
"(",
"request",
",",
"extensions",
".",
"descriptors",
")"
] | apply request extensions to an instance of :class:pyramid . | train | false |
35,719 | @register.filter
def fix_jsi18n(extrahead):
return mark_safe(unicode(extrahead).replace(u'../../../jsi18n/', reverse('admin:jsi18n')))
| [
"@",
"register",
".",
"filter",
"def",
"fix_jsi18n",
"(",
"extrahead",
")",
":",
"return",
"mark_safe",
"(",
"unicode",
"(",
"extrahead",
")",
".",
"replace",
"(",
"u'../../../jsi18n/'",
",",
"reverse",
"(",
"'admin:jsi18n'",
")",
")",
")"
] | hack to rewrite out the jsi18n script tag from an inherited admin template . | train | false |
35,721 | @tempdir
def test_full():
_system_call('virtualenv', '--no-site-packages', '.')
_system_call('bin/easy_install', '-q', 'distribute==dev')
_system_call('bin/easy_install', '-qU', 'distribute==dev')
_system_call('bin/easy_install', '-q', 'pip')
_system_call('bin/pip', 'install', '-q', 'zc.buildout')
with open('buildout.cfg', 'w') as f:
f.write(SIMPLE_BUILDOUT)
with open('bootstrap.py', 'w') as f:
f.write(urlopen(BOOTSTRAP).read())
_system_call('bin/python', 'bootstrap.py', '--distribute')
_system_call('bin/buildout', '-q')
eggs = os.listdir('eggs')
eggs.sort()
assert (len(eggs) == 3)
assert eggs[0].startswith('distribute')
assert (eggs[1:] == ['extensions-0.3-py2.6.egg', 'zc.recipe.egg-1.2.2-py2.6.egg'])
| [
"@",
"tempdir",
"def",
"test_full",
"(",
")",
":",
"_system_call",
"(",
"'virtualenv'",
",",
"'--no-site-packages'",
",",
"'.'",
")",
"_system_call",
"(",
"'bin/easy_install'",
",",
"'-q'",
",",
"'distribute==dev'",
")",
"_system_call",
"(",
"'bin/easy_install'",
",",
"'-qU'",
",",
"'distribute==dev'",
")",
"_system_call",
"(",
"'bin/easy_install'",
",",
"'-q'",
",",
"'pip'",
")",
"_system_call",
"(",
"'bin/pip'",
",",
"'install'",
",",
"'-q'",
",",
"'zc.buildout'",
")",
"with",
"open",
"(",
"'buildout.cfg'",
",",
"'w'",
")",
"as",
"f",
":",
"f",
".",
"write",
"(",
"SIMPLE_BUILDOUT",
")",
"with",
"open",
"(",
"'bootstrap.py'",
",",
"'w'",
")",
"as",
"f",
":",
"f",
".",
"write",
"(",
"urlopen",
"(",
"BOOTSTRAP",
")",
".",
"read",
"(",
")",
")",
"_system_call",
"(",
"'bin/python'",
",",
"'bootstrap.py'",
",",
"'--distribute'",
")",
"_system_call",
"(",
"'bin/buildout'",
",",
"'-q'",
")",
"eggs",
"=",
"os",
".",
"listdir",
"(",
"'eggs'",
")",
"eggs",
".",
"sort",
"(",
")",
"assert",
"(",
"len",
"(",
"eggs",
")",
"==",
"3",
")",
"assert",
"eggs",
"[",
"0",
"]",
".",
"startswith",
"(",
"'distribute'",
")",
"assert",
"(",
"eggs",
"[",
"1",
":",
"]",
"==",
"[",
"'extensions-0.3-py2.6.egg'",
",",
"'zc.recipe.egg-1.2.2-py2.6.egg'",
"]",
")"
] | virtualenv + pip + buildout . | train | false |
35,722 | def seq_minibatches(inputs, targets, batch_size, seq_length, stride=1):
assert (len(inputs) == len(targets))
n_loads = ((batch_size * stride) + (seq_length - stride))
for start_idx in range(0, ((len(inputs) - n_loads) + 1), (batch_size * stride)):
seq_inputs = np.zeros(((batch_size, seq_length) + inputs.shape[1:]), dtype=inputs.dtype)
seq_targets = np.zeros(((batch_size, seq_length) + targets.shape[1:]), dtype=targets.dtype)
for b_idx in xrange(batch_size):
start_seq_idx = (start_idx + (b_idx * stride))
end_seq_idx = (start_seq_idx + seq_length)
seq_inputs[b_idx] = inputs[start_seq_idx:end_seq_idx]
seq_targets[b_idx] = targets[start_seq_idx:end_seq_idx]
flatten_inputs = seq_inputs.reshape((((-1),) + inputs.shape[1:]))
flatten_targets = seq_targets.reshape((((-1),) + targets.shape[1:]))
(yield (flatten_inputs, flatten_targets))
| [
"def",
"seq_minibatches",
"(",
"inputs",
",",
"targets",
",",
"batch_size",
",",
"seq_length",
",",
"stride",
"=",
"1",
")",
":",
"assert",
"(",
"len",
"(",
"inputs",
")",
"==",
"len",
"(",
"targets",
")",
")",
"n_loads",
"=",
"(",
"(",
"batch_size",
"*",
"stride",
")",
"+",
"(",
"seq_length",
"-",
"stride",
")",
")",
"for",
"start_idx",
"in",
"range",
"(",
"0",
",",
"(",
"(",
"len",
"(",
"inputs",
")",
"-",
"n_loads",
")",
"+",
"1",
")",
",",
"(",
"batch_size",
"*",
"stride",
")",
")",
":",
"seq_inputs",
"=",
"np",
".",
"zeros",
"(",
"(",
"(",
"batch_size",
",",
"seq_length",
")",
"+",
"inputs",
".",
"shape",
"[",
"1",
":",
"]",
")",
",",
"dtype",
"=",
"inputs",
".",
"dtype",
")",
"seq_targets",
"=",
"np",
".",
"zeros",
"(",
"(",
"(",
"batch_size",
",",
"seq_length",
")",
"+",
"targets",
".",
"shape",
"[",
"1",
":",
"]",
")",
",",
"dtype",
"=",
"targets",
".",
"dtype",
")",
"for",
"b_idx",
"in",
"xrange",
"(",
"batch_size",
")",
":",
"start_seq_idx",
"=",
"(",
"start_idx",
"+",
"(",
"b_idx",
"*",
"stride",
")",
")",
"end_seq_idx",
"=",
"(",
"start_seq_idx",
"+",
"seq_length",
")",
"seq_inputs",
"[",
"b_idx",
"]",
"=",
"inputs",
"[",
"start_seq_idx",
":",
"end_seq_idx",
"]",
"seq_targets",
"[",
"b_idx",
"]",
"=",
"targets",
"[",
"start_seq_idx",
":",
"end_seq_idx",
"]",
"flatten_inputs",
"=",
"seq_inputs",
".",
"reshape",
"(",
"(",
"(",
"(",
"-",
"1",
")",
",",
")",
"+",
"inputs",
".",
"shape",
"[",
"1",
":",
"]",
")",
")",
"flatten_targets",
"=",
"seq_targets",
".",
"reshape",
"(",
"(",
"(",
"(",
"-",
"1",
")",
",",
")",
"+",
"targets",
".",
"shape",
"[",
"1",
":",
"]",
")",
")",
"(",
"yield",
"(",
"flatten_inputs",
",",
"flatten_targets",
")",
")"
] | generate a generator that return a batch of sequence inputs and targets . | train | true |
35,724 | def change_file_creation_mask(mask):
try:
os.umask(mask)
except Exception as exc:
error = DaemonOSEnvironmentError(('Unable to change file creation mask (%(exc)s)' % vars()))
raise error
| [
"def",
"change_file_creation_mask",
"(",
"mask",
")",
":",
"try",
":",
"os",
".",
"umask",
"(",
"mask",
")",
"except",
"Exception",
"as",
"exc",
":",
"error",
"=",
"DaemonOSEnvironmentError",
"(",
"(",
"'Unable to change file creation mask (%(exc)s)'",
"%",
"vars",
"(",
")",
")",
")",
"raise",
"error"
] | change the file creation mask for this process . | train | false |
35,726 | def write_(path, key, value, profile=None):
result = _query('POST', path, profile=profile, data=json.dumps({key: value}))
return read_(path, key, profile)
| [
"def",
"write_",
"(",
"path",
",",
"key",
",",
"value",
",",
"profile",
"=",
"None",
")",
":",
"result",
"=",
"_query",
"(",
"'POST'",
",",
"path",
",",
"profile",
"=",
"profile",
",",
"data",
"=",
"json",
".",
"dumps",
"(",
"{",
"key",
":",
"value",
"}",
")",
")",
"return",
"read_",
"(",
"path",
",",
"key",
",",
"profile",
")"
] | set a key/value pair in the vault service . | train | false |
35,728 | def endmsg():
print " [*] Your payload files have been generated, don't get caught!"
print helpers.color(" [!] And don't submit samples to any online scanner! ;)\n", yellow=True)
| [
"def",
"endmsg",
"(",
")",
":",
"print",
"\" [*] Your payload files have been generated, don't get caught!\"",
"print",
"helpers",
".",
"color",
"(",
"\" [!] And don't submit samples to any online scanner! ;)\\n\"",
",",
"yellow",
"=",
"True",
")"
] | print the exit message . | train | false |
35,729 | def retrieve_auth():
def get_kwargs(envargs):
return dict([(arg.kwarg, os.environ.get(arg.envarg)) for arg in envargs])
mc_kwargs = get_kwargs(mc_envargs)
mm_kwargs = get_kwargs(mm_envargs)
if (not all([mc_kwargs[arg] for arg in (u'email', u'password', u'android_id')])):
if os.environ.get(u'TRAVIS'):
print(u'on Travis but could not read auth from environ; quitting.')
sys.exit(1)
mc_kwargs.update(zip([u'email', u'password', u'android_id'], prompt_for_mc_auth()))
if (mm_kwargs[u'oauth_credentials'] is None):
if (not os.path.isfile(OAUTH_FILEPATH)):
raise ValueError(u'You must have oauth credentials stored at the default path by Musicmanager.perform_oauth prior to running.')
del mm_kwargs[u'oauth_credentials']
else:
mm_kwargs[u'oauth_credentials'] = credentials_from_refresh_token(mm_kwargs[u'oauth_credentials'])
return (mc_kwargs, mm_kwargs)
| [
"def",
"retrieve_auth",
"(",
")",
":",
"def",
"get_kwargs",
"(",
"envargs",
")",
":",
"return",
"dict",
"(",
"[",
"(",
"arg",
".",
"kwarg",
",",
"os",
".",
"environ",
".",
"get",
"(",
"arg",
".",
"envarg",
")",
")",
"for",
"arg",
"in",
"envargs",
"]",
")",
"mc_kwargs",
"=",
"get_kwargs",
"(",
"mc_envargs",
")",
"mm_kwargs",
"=",
"get_kwargs",
"(",
"mm_envargs",
")",
"if",
"(",
"not",
"all",
"(",
"[",
"mc_kwargs",
"[",
"arg",
"]",
"for",
"arg",
"in",
"(",
"u'email'",
",",
"u'password'",
",",
"u'android_id'",
")",
"]",
")",
")",
":",
"if",
"os",
".",
"environ",
".",
"get",
"(",
"u'TRAVIS'",
")",
":",
"print",
"(",
"u'on Travis but could not read auth from environ; quitting.'",
")",
"sys",
".",
"exit",
"(",
"1",
")",
"mc_kwargs",
".",
"update",
"(",
"zip",
"(",
"[",
"u'email'",
",",
"u'password'",
",",
"u'android_id'",
"]",
",",
"prompt_for_mc_auth",
"(",
")",
")",
")",
"if",
"(",
"mm_kwargs",
"[",
"u'oauth_credentials'",
"]",
"is",
"None",
")",
":",
"if",
"(",
"not",
"os",
".",
"path",
".",
"isfile",
"(",
"OAUTH_FILEPATH",
")",
")",
":",
"raise",
"ValueError",
"(",
"u'You must have oauth credentials stored at the default path by Musicmanager.perform_oauth prior to running.'",
")",
"del",
"mm_kwargs",
"[",
"u'oauth_credentials'",
"]",
"else",
":",
"mm_kwargs",
"[",
"u'oauth_credentials'",
"]",
"=",
"credentials_from_refresh_token",
"(",
"mm_kwargs",
"[",
"u'oauth_credentials'",
"]",
")",
"return",
"(",
"mc_kwargs",
",",
"mm_kwargs",
")"
] | searches the env for auth . | train | false |
35,732 | @pytest.mark.skipif('no_real_s3_credentials()')
def test_301_redirect():
import boto.s3.connection
bucket_name = bucket_name_mangle('wal-e-test-301-redirect')
with pytest.raises(boto.exception.S3ResponseError) as e:
with FreshBucket(bucket_name, calling_format=OrdinaryCallingFormat()) as fb:
fb.create(location='us-west-1')
assert (e.value.status == 301)
| [
"@",
"pytest",
".",
"mark",
".",
"skipif",
"(",
"'no_real_s3_credentials()'",
")",
"def",
"test_301_redirect",
"(",
")",
":",
"import",
"boto",
".",
"s3",
".",
"connection",
"bucket_name",
"=",
"bucket_name_mangle",
"(",
"'wal-e-test-301-redirect'",
")",
"with",
"pytest",
".",
"raises",
"(",
"boto",
".",
"exception",
".",
"S3ResponseError",
")",
"as",
"e",
":",
"with",
"FreshBucket",
"(",
"bucket_name",
",",
"calling_format",
"=",
"OrdinaryCallingFormat",
"(",
")",
")",
"as",
"fb",
":",
"fb",
".",
"create",
"(",
"location",
"=",
"'us-west-1'",
")",
"assert",
"(",
"e",
".",
"value",
".",
"status",
"==",
"301",
")"
] | integration test for bucket naming issues this test . | train | false |
35,734 | def nti(s):
if (s[0] != chr(128)):
try:
n = int((nts(s) or '0'), 8)
except ValueError:
raise HeaderError('invalid header')
else:
n = 0L
for i in xrange((len(s) - 1)):
n <<= 8
n += ord(s[(i + 1)])
return n
| [
"def",
"nti",
"(",
"s",
")",
":",
"if",
"(",
"s",
"[",
"0",
"]",
"!=",
"chr",
"(",
"128",
")",
")",
":",
"try",
":",
"n",
"=",
"int",
"(",
"(",
"nts",
"(",
"s",
")",
"or",
"'0'",
")",
",",
"8",
")",
"except",
"ValueError",
":",
"raise",
"HeaderError",
"(",
"'invalid header'",
")",
"else",
":",
"n",
"=",
"0",
"L",
"for",
"i",
"in",
"xrange",
"(",
"(",
"len",
"(",
"s",
")",
"-",
"1",
")",
")",
":",
"n",
"<<=",
"8",
"n",
"+=",
"ord",
"(",
"s",
"[",
"(",
"i",
"+",
"1",
")",
"]",
")",
"return",
"n"
] | convert a number field to a python number . | train | false |
35,735 | def fourier_shift(input, shift, n=(-1), axis=(-1), output=None):
input = numpy.asarray(input)
(output, return_value) = _get_output_fourier_complex(output, input)
axis = _ni_support._check_axis(axis, input.ndim)
shifts = _ni_support._normalize_sequence(shift, input.ndim)
shifts = numpy.asarray(shifts, dtype=numpy.float64)
if (not shifts.flags.contiguous):
shifts = shifts.copy()
_nd_image.fourier_shift(input, shifts, n, axis, output)
return return_value
| [
"def",
"fourier_shift",
"(",
"input",
",",
"shift",
",",
"n",
"=",
"(",
"-",
"1",
")",
",",
"axis",
"=",
"(",
"-",
"1",
")",
",",
"output",
"=",
"None",
")",
":",
"input",
"=",
"numpy",
".",
"asarray",
"(",
"input",
")",
"(",
"output",
",",
"return_value",
")",
"=",
"_get_output_fourier_complex",
"(",
"output",
",",
"input",
")",
"axis",
"=",
"_ni_support",
".",
"_check_axis",
"(",
"axis",
",",
"input",
".",
"ndim",
")",
"shifts",
"=",
"_ni_support",
".",
"_normalize_sequence",
"(",
"shift",
",",
"input",
".",
"ndim",
")",
"shifts",
"=",
"numpy",
".",
"asarray",
"(",
"shifts",
",",
"dtype",
"=",
"numpy",
".",
"float64",
")",
"if",
"(",
"not",
"shifts",
".",
"flags",
".",
"contiguous",
")",
":",
"shifts",
"=",
"shifts",
".",
"copy",
"(",
")",
"_nd_image",
".",
"fourier_shift",
"(",
"input",
",",
"shifts",
",",
"n",
",",
"axis",
",",
"output",
")",
"return",
"return_value"
] | multi-dimensional fourier shift filter . | train | false |
35,738 | def popValue():
return getCurrentThreadData().valueStack.pop()
| [
"def",
"popValue",
"(",
")",
":",
"return",
"getCurrentThreadData",
"(",
")",
".",
"valueStack",
".",
"pop",
"(",
")"
] | pop value from the stack . | train | false |
35,740 | def _children_with_tag(element, tag):
return itertools.chain(element.iterfind((_OLD_NAMESPACE_PREFIX + tag)), element.iterfind((_NEW_NAMESPACE_PREFIX + tag)))
| [
"def",
"_children_with_tag",
"(",
"element",
",",
"tag",
")",
":",
"return",
"itertools",
".",
"chain",
"(",
"element",
".",
"iterfind",
"(",
"(",
"_OLD_NAMESPACE_PREFIX",
"+",
"tag",
")",
")",
",",
"element",
".",
"iterfind",
"(",
"(",
"_NEW_NAMESPACE_PREFIX",
"+",
"tag",
")",
")",
")"
] | returns child elements of the given element with the given tag . | train | false |
35,742 | def format_time(seconds, count=3, accuracy=6, simple=False):
if simple:
periods = [('c', ((((60 * 60) * 24) * 365) * 100)), ('de', ((((60 * 60) * 24) * 365) * 10)), ('y', (((60 * 60) * 24) * 365)), ('mo', (((60 * 60) * 24) * 30)), ('d', ((60 * 60) * 24)), ('h', (60 * 60)), ('m', 60), ('s', 1)]
else:
periods = [(('century', 'centuries'), ((((60 * 60) * 24) * 365) * 100)), (('decade', 'decades'), ((((60 * 60) * 24) * 365) * 10)), (('year', 'years'), (((60 * 60) * 24) * 365)), (('month', 'months'), (((60 * 60) * 24) * 30)), (('day', 'days'), ((60 * 60) * 24)), (('hour', 'hours'), (60 * 60)), (('minute', 'minutes'), 60), (('second', 'seconds'), 1)]
periods = periods[(- accuracy):]
strings = []
i = 0
for (period_name, period_seconds) in periods:
if (i < count):
if (seconds > period_seconds):
(period_value, seconds) = divmod(seconds, period_seconds)
i += 1
if simple:
strings.append('{}{}'.format(period_value, period_name))
elif (period_value == 1):
strings.append('{} {}'.format(period_value, period_name[0]))
else:
strings.append('{} {}'.format(period_value, period_name[1]))
else:
break
if simple:
return ' '.join(strings)
else:
return formatting.get_text_list(strings, 'and')
| [
"def",
"format_time",
"(",
"seconds",
",",
"count",
"=",
"3",
",",
"accuracy",
"=",
"6",
",",
"simple",
"=",
"False",
")",
":",
"if",
"simple",
":",
"periods",
"=",
"[",
"(",
"'c'",
",",
"(",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"365",
")",
"*",
"100",
")",
")",
",",
"(",
"'de'",
",",
"(",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"365",
")",
"*",
"10",
")",
")",
",",
"(",
"'y'",
",",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"365",
")",
")",
",",
"(",
"'mo'",
",",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"30",
")",
")",
",",
"(",
"'d'",
",",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
")",
",",
"(",
"'h'",
",",
"(",
"60",
"*",
"60",
")",
")",
",",
"(",
"'m'",
",",
"60",
")",
",",
"(",
"'s'",
",",
"1",
")",
"]",
"else",
":",
"periods",
"=",
"[",
"(",
"(",
"'century'",
",",
"'centuries'",
")",
",",
"(",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"365",
")",
"*",
"100",
")",
")",
",",
"(",
"(",
"'decade'",
",",
"'decades'",
")",
",",
"(",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"365",
")",
"*",
"10",
")",
")",
",",
"(",
"(",
"'year'",
",",
"'years'",
")",
",",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"365",
")",
")",
",",
"(",
"(",
"'month'",
",",
"'months'",
")",
",",
"(",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
"*",
"30",
")",
")",
",",
"(",
"(",
"'day'",
",",
"'days'",
")",
",",
"(",
"(",
"60",
"*",
"60",
")",
"*",
"24",
")",
")",
",",
"(",
"(",
"'hour'",
",",
"'hours'",
")",
",",
"(",
"60",
"*",
"60",
")",
")",
",",
"(",
"(",
"'minute'",
",",
"'minutes'",
")",
",",
"60",
")",
",",
"(",
"(",
"'second'",
",",
"'seconds'",
")",
",",
"1",
")",
"]",
"periods",
"=",
"periods",
"[",
"(",
"-",
"accuracy",
")",
":",
"]",
"strings",
"=",
"[",
"]",
"i",
"=",
"0",
"for",
"(",
"period_name",
",",
"period_seconds",
")",
"in",
"periods",
":",
"if",
"(",
"i",
"<",
"count",
")",
":",
"if",
"(",
"seconds",
">",
"period_seconds",
")",
":",
"(",
"period_value",
",",
"seconds",
")",
"=",
"divmod",
"(",
"seconds",
",",
"period_seconds",
")",
"i",
"+=",
"1",
"if",
"simple",
":",
"strings",
".",
"append",
"(",
"'{}{}'",
".",
"format",
"(",
"period_value",
",",
"period_name",
")",
")",
"elif",
"(",
"period_value",
"==",
"1",
")",
":",
"strings",
".",
"append",
"(",
"'{} {}'",
".",
"format",
"(",
"period_value",
",",
"period_name",
"[",
"0",
"]",
")",
")",
"else",
":",
"strings",
".",
"append",
"(",
"'{} {}'",
".",
"format",
"(",
"period_value",
",",
"period_name",
"[",
"1",
"]",
")",
")",
"else",
":",
"break",
"if",
"simple",
":",
"return",
"' '",
".",
"join",
"(",
"strings",
")",
"else",
":",
"return",
"formatting",
".",
"get_text_list",
"(",
"strings",
",",
"'and'",
")"
] | returns a timestamp formatted from given time using separators . | train | false |
35,744 | def _setLayerRule(layer, rule=None):
layer.ClearField('include')
layer.ClearField('exclude')
if (rule is not None):
layer.include.add().CopyFrom(rule)
| [
"def",
"_setLayerRule",
"(",
"layer",
",",
"rule",
"=",
"None",
")",
":",
"layer",
".",
"ClearField",
"(",
"'include'",
")",
"layer",
".",
"ClearField",
"(",
"'exclude'",
")",
"if",
"(",
"rule",
"is",
"not",
"None",
")",
":",
"layer",
".",
"include",
".",
"add",
"(",
")",
".",
"CopyFrom",
"(",
"rule",
")"
] | set a new include rule for this layer if rule is none . | train | false |
35,745 | def get_module_source_path(modname, basename=None):
srcpath = get_module_path(modname)
parentdir = osp.join(srcpath, osp.pardir)
if osp.isfile(parentdir):
srcpath = osp.abspath(osp.join(osp.join(parentdir, osp.pardir), modname))
if (basename is not None):
srcpath = osp.abspath(osp.join(srcpath, basename))
return srcpath
| [
"def",
"get_module_source_path",
"(",
"modname",
",",
"basename",
"=",
"None",
")",
":",
"srcpath",
"=",
"get_module_path",
"(",
"modname",
")",
"parentdir",
"=",
"osp",
".",
"join",
"(",
"srcpath",
",",
"osp",
".",
"pardir",
")",
"if",
"osp",
".",
"isfile",
"(",
"parentdir",
")",
":",
"srcpath",
"=",
"osp",
".",
"abspath",
"(",
"osp",
".",
"join",
"(",
"osp",
".",
"join",
"(",
"parentdir",
",",
"osp",
".",
"pardir",
")",
",",
"modname",
")",
")",
"if",
"(",
"basename",
"is",
"not",
"None",
")",
":",
"srcpath",
"=",
"osp",
".",
"abspath",
"(",
"osp",
".",
"join",
"(",
"srcpath",
",",
"basename",
")",
")",
"return",
"srcpath"
] | return module *modname* source path if *basename* is specified . | train | true |
35,746 | def to_native_str(text, encoding=None, errors='strict'):
if six.PY2:
return to_bytes(text, encoding, errors)
else:
return to_unicode(text, encoding, errors)
| [
"def",
"to_native_str",
"(",
"text",
",",
"encoding",
"=",
"None",
",",
"errors",
"=",
"'strict'",
")",
":",
"if",
"six",
".",
"PY2",
":",
"return",
"to_bytes",
"(",
"text",
",",
"encoding",
",",
"errors",
")",
"else",
":",
"return",
"to_unicode",
"(",
"text",
",",
"encoding",
",",
"errors",
")"
] | return str representation of text . | train | false |
35,747 | def json_load(filename):
with open(filename, u'r') as fh:
return json.load(fh, object_hook=_json_decode)
| [
"def",
"json_load",
"(",
"filename",
")",
":",
"with",
"open",
"(",
"filename",
",",
"u'r'",
")",
"as",
"fh",
":",
"return",
"json",
".",
"load",
"(",
"fh",
",",
"object_hook",
"=",
"_json_decode",
")"
] | loads a data structure as json from the named file . | train | false |
35,748 | def secho(message=None, file=None, nl=True, err=False, color=None, **styles):
if (message is not None):
message = style(message, **styles)
return echo(message, file=file, nl=nl, err=err, color=color)
| [
"def",
"secho",
"(",
"message",
"=",
"None",
",",
"file",
"=",
"None",
",",
"nl",
"=",
"True",
",",
"err",
"=",
"False",
",",
"color",
"=",
"None",
",",
"**",
"styles",
")",
":",
"if",
"(",
"message",
"is",
"not",
"None",
")",
":",
"message",
"=",
"style",
"(",
"message",
",",
"**",
"styles",
")",
"return",
"echo",
"(",
"message",
",",
"file",
"=",
"file",
",",
"nl",
"=",
"nl",
",",
"err",
"=",
"err",
",",
"color",
"=",
"color",
")"
] | this function combines :func:echo and :func:style into one call . | train | true |
35,749 | def write_mergecells(worksheet):
cells = worksheet._merged_cells
if (not cells):
return
merge = Element('mergeCells', count=('%d' % len(cells)))
for range_string in cells:
merge.append(Element('mergeCell', ref=range_string))
return merge
| [
"def",
"write_mergecells",
"(",
"worksheet",
")",
":",
"cells",
"=",
"worksheet",
".",
"_merged_cells",
"if",
"(",
"not",
"cells",
")",
":",
"return",
"merge",
"=",
"Element",
"(",
"'mergeCells'",
",",
"count",
"=",
"(",
"'%d'",
"%",
"len",
"(",
"cells",
")",
")",
")",
"for",
"range_string",
"in",
"cells",
":",
"merge",
".",
"append",
"(",
"Element",
"(",
"'mergeCell'",
",",
"ref",
"=",
"range_string",
")",
")",
"return",
"merge"
] | write merged cells to xml . | train | false |
35,752 | def ULTOSC(barDs, count, timeperiod1=(- (2 ** 31)), timeperiod2=(- (2 ** 31)), timeperiod3=(- (2 ** 31))):
return call_talib_with_hlc(barDs, count, talib.ULTOSC, timeperiod1, timeperiod2, timeperiod3)
| [
"def",
"ULTOSC",
"(",
"barDs",
",",
"count",
",",
"timeperiod1",
"=",
"(",
"-",
"(",
"2",
"**",
"31",
")",
")",
",",
"timeperiod2",
"=",
"(",
"-",
"(",
"2",
"**",
"31",
")",
")",
",",
"timeperiod3",
"=",
"(",
"-",
"(",
"2",
"**",
"31",
")",
")",
")",
":",
"return",
"call_talib_with_hlc",
"(",
"barDs",
",",
"count",
",",
"talib",
".",
"ULTOSC",
",",
"timeperiod1",
",",
"timeperiod2",
",",
"timeperiod3",
")"
] | ultimate oscillator . | train | false |
35,755 | def network_get_all_by_instance(context, instance_id):
return IMPL.network_get_all_by_instance(context, instance_id)
| [
"def",
"network_get_all_by_instance",
"(",
"context",
",",
"instance_id",
")",
":",
"return",
"IMPL",
".",
"network_get_all_by_instance",
"(",
"context",
",",
"instance_id",
")"
] | get all networks by instance id or raise if none exist . | train | false |
35,756 | @register.inclusion_tag('inclusion.html', takes_context=True)
def inclusion_params_and_context(context, arg):
return {'result': ('inclusion_params_and_context - Expected result (context value: %s): %s' % (context['value'], arg))}
| [
"@",
"register",
".",
"inclusion_tag",
"(",
"'inclusion.html'",
",",
"takes_context",
"=",
"True",
")",
"def",
"inclusion_params_and_context",
"(",
"context",
",",
"arg",
")",
":",
"return",
"{",
"'result'",
":",
"(",
"'inclusion_params_and_context - Expected result (context value: %s): %s'",
"%",
"(",
"context",
"[",
"'value'",
"]",
",",
"arg",
")",
")",
"}"
] | expected inclusion_params_and_context __doc__ . | train | false |
35,757 | def _remove_persistent_module(mod, comment):
conf = _get_modules_conf()
mod_name = _strip_module_name(mod)
if ((not mod_name) or (mod_name not in mod_list(True))):
return set()
escape_mod = re.escape(mod)
if comment:
__salt__['file.comment'](conf, '^[ DCTB ]*{0}[ DCTB ]?'.format(escape_mod))
else:
__salt__['file.sed'](conf, '^[ DCTB ]*{0}[ DCTB ]?'.format(escape_mod), '')
return set([mod_name])
| [
"def",
"_remove_persistent_module",
"(",
"mod",
",",
"comment",
")",
":",
"conf",
"=",
"_get_modules_conf",
"(",
")",
"mod_name",
"=",
"_strip_module_name",
"(",
"mod",
")",
"if",
"(",
"(",
"not",
"mod_name",
")",
"or",
"(",
"mod_name",
"not",
"in",
"mod_list",
"(",
"True",
")",
")",
")",
":",
"return",
"set",
"(",
")",
"escape_mod",
"=",
"re",
".",
"escape",
"(",
"mod",
")",
"if",
"comment",
":",
"__salt__",
"[",
"'file.comment'",
"]",
"(",
"conf",
",",
"'^[ DCTB ]*{0}[ DCTB ]?'",
".",
"format",
"(",
"escape_mod",
")",
")",
"else",
":",
"__salt__",
"[",
"'file.sed'",
"]",
"(",
"conf",
",",
"'^[ DCTB ]*{0}[ DCTB ]?'",
".",
"format",
"(",
"escape_mod",
")",
",",
"''",
")",
"return",
"set",
"(",
"[",
"mod_name",
"]",
")"
] | remove module from configuration file . | train | true |
35,758 | def _generate_storage_paths(app_id):
if (sys.platform == 'win32'):
user_format = ''
else:
try:
user_name = getpass.getuser()
except Exception:
user_format = ''
else:
user_format = ('.%s' % user_name)
tempdir = tempfile.gettempdir()
(yield os.path.join(tempdir, ('appengine.%s%s' % (app_id, user_format))))
for i in itertools.count(1):
(yield os.path.join(tempdir, ('appengine.%s%s.%d' % (app_id, user_format, i))))
| [
"def",
"_generate_storage_paths",
"(",
"app_id",
")",
":",
"if",
"(",
"sys",
".",
"platform",
"==",
"'win32'",
")",
":",
"user_format",
"=",
"''",
"else",
":",
"try",
":",
"user_name",
"=",
"getpass",
".",
"getuser",
"(",
")",
"except",
"Exception",
":",
"user_format",
"=",
"''",
"else",
":",
"user_format",
"=",
"(",
"'.%s'",
"%",
"user_name",
")",
"tempdir",
"=",
"tempfile",
".",
"gettempdir",
"(",
")",
"(",
"yield",
"os",
".",
"path",
".",
"join",
"(",
"tempdir",
",",
"(",
"'appengine.%s%s'",
"%",
"(",
"app_id",
",",
"user_format",
")",
")",
")",
")",
"for",
"i",
"in",
"itertools",
".",
"count",
"(",
"1",
")",
":",
"(",
"yield",
"os",
".",
"path",
".",
"join",
"(",
"tempdir",
",",
"(",
"'appengine.%s%s.%d'",
"%",
"(",
"app_id",
",",
"user_format",
",",
"i",
")",
")",
")",
")"
] | yield an infinite sequence of possible storage paths . | train | false |
35,759 | def get_sw_version(strip_build_num=False):
try:
from freenasOS import Configuration
except ImportError:
Configuration = None
global _VERSION
if (_VERSION is None):
if Configuration:
conf = Configuration.Configuration()
sys_mani = conf.SystemManifest()
if sys_mani:
_VERSION = sys_mani.Sequence()
if (_VERSION is None):
with open(VERSION_FILE) as fd:
_VERSION = fd.read().strip()
if strip_build_num:
return _VERSION.split(' ')[0]
return _VERSION
| [
"def",
"get_sw_version",
"(",
"strip_build_num",
"=",
"False",
")",
":",
"try",
":",
"from",
"freenasOS",
"import",
"Configuration",
"except",
"ImportError",
":",
"Configuration",
"=",
"None",
"global",
"_VERSION",
"if",
"(",
"_VERSION",
"is",
"None",
")",
":",
"if",
"Configuration",
":",
"conf",
"=",
"Configuration",
".",
"Configuration",
"(",
")",
"sys_mani",
"=",
"conf",
".",
"SystemManifest",
"(",
")",
"if",
"sys_mani",
":",
"_VERSION",
"=",
"sys_mani",
".",
"Sequence",
"(",
")",
"if",
"(",
"_VERSION",
"is",
"None",
")",
":",
"with",
"open",
"(",
"VERSION_FILE",
")",
"as",
"fd",
":",
"_VERSION",
"=",
"fd",
".",
"read",
"(",
")",
".",
"strip",
"(",
")",
"if",
"strip_build_num",
":",
"return",
"_VERSION",
".",
"split",
"(",
"' '",
")",
"[",
"0",
"]",
"return",
"_VERSION"
] | return the full version string . | train | false |
35,760 | def _float_ones_like(x):
dtype = x.type.dtype
if (dtype not in tensor.float_dtypes):
dtype = theano.config.floatX
return tensor.ones_like(x, dtype=dtype)
| [
"def",
"_float_ones_like",
"(",
"x",
")",
":",
"dtype",
"=",
"x",
".",
"type",
".",
"dtype",
"if",
"(",
"dtype",
"not",
"in",
"tensor",
".",
"float_dtypes",
")",
":",
"dtype",
"=",
"theano",
".",
"config",
".",
"floatX",
"return",
"tensor",
".",
"ones_like",
"(",
"x",
",",
"dtype",
"=",
"dtype",
")"
] | like ones_like . | train | false |
35,761 | def py_factorial(n):
inner = outer = 1
for i in reversed(range(n.bit_length())):
inner *= partial_product((((n >> (i + 1)) + 1) | 1), (((n >> i) + 1) | 1))
outer *= inner
return (outer << (n - count_set_bits(n)))
| [
"def",
"py_factorial",
"(",
"n",
")",
":",
"inner",
"=",
"outer",
"=",
"1",
"for",
"i",
"in",
"reversed",
"(",
"range",
"(",
"n",
".",
"bit_length",
"(",
")",
")",
")",
":",
"inner",
"*=",
"partial_product",
"(",
"(",
"(",
"(",
"n",
">>",
"(",
"i",
"+",
"1",
")",
")",
"+",
"1",
")",
"|",
"1",
")",
",",
"(",
"(",
"(",
"n",
">>",
"i",
")",
"+",
"1",
")",
"|",
"1",
")",
")",
"outer",
"*=",
"inner",
"return",
"(",
"outer",
"<<",
"(",
"n",
"-",
"count_set_bits",
"(",
"n",
")",
")",
")"
] | factorial of nonnegative integer n . | train | false |
35,762 | def _is_match(tagged_phrase, cfg):
copy = list(tagged_phrase)
merge = True
while merge:
merge = False
for i in range((len(copy) - 1)):
(first, second) = (copy[i], copy[(i + 1)])
key = (first[1], second[1])
value = cfg.get(key, None)
if value:
merge = True
copy.pop(i)
copy.pop(i)
match = u'{0} {1}'.format(first[0], second[0])
pos = value
copy.insert(i, (match, pos))
break
match = any([(t[1] in (u'NNP', u'NNI')) for t in copy])
return match
| [
"def",
"_is_match",
"(",
"tagged_phrase",
",",
"cfg",
")",
":",
"copy",
"=",
"list",
"(",
"tagged_phrase",
")",
"merge",
"=",
"True",
"while",
"merge",
":",
"merge",
"=",
"False",
"for",
"i",
"in",
"range",
"(",
"(",
"len",
"(",
"copy",
")",
"-",
"1",
")",
")",
":",
"(",
"first",
",",
"second",
")",
"=",
"(",
"copy",
"[",
"i",
"]",
",",
"copy",
"[",
"(",
"i",
"+",
"1",
")",
"]",
")",
"key",
"=",
"(",
"first",
"[",
"1",
"]",
",",
"second",
"[",
"1",
"]",
")",
"value",
"=",
"cfg",
".",
"get",
"(",
"key",
",",
"None",
")",
"if",
"value",
":",
"merge",
"=",
"True",
"copy",
".",
"pop",
"(",
"i",
")",
"copy",
".",
"pop",
"(",
"i",
")",
"match",
"=",
"u'{0} {1}'",
".",
"format",
"(",
"first",
"[",
"0",
"]",
",",
"second",
"[",
"0",
"]",
")",
"pos",
"=",
"value",
"copy",
".",
"insert",
"(",
"i",
",",
"(",
"match",
",",
"pos",
")",
")",
"break",
"match",
"=",
"any",
"(",
"[",
"(",
"t",
"[",
"1",
"]",
"in",
"(",
"u'NNP'",
",",
"u'NNI'",
")",
")",
"for",
"t",
"in",
"copy",
"]",
")",
"return",
"match"
] | return whether or not a tagged phrases matches a context-free grammar . | train | false |
35,763 | def make_alternating_boolean_array(shape, first_value=True):
if (len(shape) != 2):
raise ValueError('Shape must be 2-dimensional. Given shape was {}'.format(shape))
alternating = np.empty(shape, dtype=np.bool)
for row in alternating:
row[::2] = first_value
row[1::2] = (not first_value)
first_value = (not first_value)
return alternating
| [
"def",
"make_alternating_boolean_array",
"(",
"shape",
",",
"first_value",
"=",
"True",
")",
":",
"if",
"(",
"len",
"(",
"shape",
")",
"!=",
"2",
")",
":",
"raise",
"ValueError",
"(",
"'Shape must be 2-dimensional. Given shape was {}'",
".",
"format",
"(",
"shape",
")",
")",
"alternating",
"=",
"np",
".",
"empty",
"(",
"shape",
",",
"dtype",
"=",
"np",
".",
"bool",
")",
"for",
"row",
"in",
"alternating",
":",
"row",
"[",
":",
":",
"2",
"]",
"=",
"first_value",
"row",
"[",
"1",
":",
":",
"2",
"]",
"=",
"(",
"not",
"first_value",
")",
"first_value",
"=",
"(",
"not",
"first_value",
")",
"return",
"alternating"
] | create a 2d numpy array with the given shape containing alternating values of false . | train | false |
35,764 | def getargspec(obj):
if inspect.isfunction(obj):
func_obj = obj
elif inspect.ismethod(obj):
func_obj = obj.__func__
else:
raise TypeError('arg is not a Python function')
(args, varargs, varkw) = inspect.getargs(func_obj.__code__)
return (args, varargs, varkw, func_obj.__defaults__)
| [
"def",
"getargspec",
"(",
"obj",
")",
":",
"if",
"inspect",
".",
"isfunction",
"(",
"obj",
")",
":",
"func_obj",
"=",
"obj",
"elif",
"inspect",
".",
"ismethod",
"(",
"obj",
")",
":",
"func_obj",
"=",
"obj",
".",
"__func__",
"else",
":",
"raise",
"TypeError",
"(",
"'arg is not a Python function'",
")",
"(",
"args",
",",
"varargs",
",",
"varkw",
")",
"=",
"inspect",
".",
"getargs",
"(",
"func_obj",
".",
"__code__",
")",
"return",
"(",
"args",
",",
"varargs",
",",
"varkw",
",",
"func_obj",
".",
"__defaults__",
")"
] | get the names and default values of a functions arguments . | train | false |
35,765 | def user_list(database=None, user=None, password=None, host=None, port=None):
client = _client(user=user, password=password, host=host, port=port)
if (not database):
return client.get_list_cluster_admins()
client.switch_database(database)
return client.get_list_users()
| [
"def",
"user_list",
"(",
"database",
"=",
"None",
",",
"user",
"=",
"None",
",",
"password",
"=",
"None",
",",
"host",
"=",
"None",
",",
"port",
"=",
"None",
")",
":",
"client",
"=",
"_client",
"(",
"user",
"=",
"user",
",",
"password",
"=",
"password",
",",
"host",
"=",
"host",
",",
"port",
"=",
"port",
")",
"if",
"(",
"not",
"database",
")",
":",
"return",
"client",
".",
"get_list_cluster_admins",
"(",
")",
"client",
".",
"switch_database",
"(",
"database",
")",
"return",
"client",
".",
"get_list_users",
"(",
")"
] | return a list of users on a mysql server cli example: . | train | true |
35,766 | def get_cycles(graph_dict, vertices=None):
if (not graph_dict):
return ()
result = []
if (vertices is None):
vertices = graph_dict.keys()
for vertice in vertices:
_get_cycles(graph_dict, [], set(), result, vertice)
return result
| [
"def",
"get_cycles",
"(",
"graph_dict",
",",
"vertices",
"=",
"None",
")",
":",
"if",
"(",
"not",
"graph_dict",
")",
":",
"return",
"(",
")",
"result",
"=",
"[",
"]",
"if",
"(",
"vertices",
"is",
"None",
")",
":",
"vertices",
"=",
"graph_dict",
".",
"keys",
"(",
")",
"for",
"vertice",
"in",
"vertices",
":",
"_get_cycles",
"(",
"graph_dict",
",",
"[",
"]",
",",
"set",
"(",
")",
",",
"result",
",",
"vertice",
")",
"return",
"result"
] | given a dictionary representing an ordered graph . | train | true |
35,767 | def get_save_notes(dire):
return read_in(os.path.join(dire, 'CHANGES_SINCE'))
| [
"def",
"get_save_notes",
"(",
"dire",
")",
":",
"return",
"read_in",
"(",
"os",
".",
"path",
".",
"join",
"(",
"dire",
",",
"'CHANGES_SINCE'",
")",
")"
] | read save notes . | train | false |
35,768 | def package_check(pkg_name, version=None, app='pandas', checker=LooseVersion, exc_failed_import=ImportError, exc_failed_check=RuntimeError):
if app:
msg = ('%s requires %s' % (app, pkg_name))
else:
msg = ('module requires %s' % pkg_name)
if version:
msg += (' with version >= %s' % (version,))
try:
mod = __import__(pkg_name)
except ImportError:
raise exc_failed_import(msg)
if (not version):
return
try:
have_version = mod.__version__
except AttributeError:
raise exc_failed_check(('Cannot find version for %s' % pkg_name))
if (checker(have_version) < checker(version)):
raise exc_failed_check(msg)
| [
"def",
"package_check",
"(",
"pkg_name",
",",
"version",
"=",
"None",
",",
"app",
"=",
"'pandas'",
",",
"checker",
"=",
"LooseVersion",
",",
"exc_failed_import",
"=",
"ImportError",
",",
"exc_failed_check",
"=",
"RuntimeError",
")",
":",
"if",
"app",
":",
"msg",
"=",
"(",
"'%s requires %s'",
"%",
"(",
"app",
",",
"pkg_name",
")",
")",
"else",
":",
"msg",
"=",
"(",
"'module requires %s'",
"%",
"pkg_name",
")",
"if",
"version",
":",
"msg",
"+=",
"(",
"' with version >= %s'",
"%",
"(",
"version",
",",
")",
")",
"try",
":",
"mod",
"=",
"__import__",
"(",
"pkg_name",
")",
"except",
"ImportError",
":",
"raise",
"exc_failed_import",
"(",
"msg",
")",
"if",
"(",
"not",
"version",
")",
":",
"return",
"try",
":",
"have_version",
"=",
"mod",
".",
"__version__",
"except",
"AttributeError",
":",
"raise",
"exc_failed_check",
"(",
"(",
"'Cannot find version for %s'",
"%",
"pkg_name",
")",
")",
"if",
"(",
"checker",
"(",
"have_version",
")",
"<",
"checker",
"(",
"version",
")",
")",
":",
"raise",
"exc_failed_check",
"(",
"msg",
")"
] | check that the minimal version of the required package is installed . | train | false |
35,770 | def _detection_limit_index(obs, cohn):
if (cohn.shape[0] > 0):
(index,) = numpy.where((cohn['lower_dl'] <= obs))
det_limit_index = index[(-1)]
else:
det_limit_index = 0
return det_limit_index
| [
"def",
"_detection_limit_index",
"(",
"obs",
",",
"cohn",
")",
":",
"if",
"(",
"cohn",
".",
"shape",
"[",
"0",
"]",
">",
"0",
")",
":",
"(",
"index",
",",
")",
"=",
"numpy",
".",
"where",
"(",
"(",
"cohn",
"[",
"'lower_dl'",
"]",
"<=",
"obs",
")",
")",
"det_limit_index",
"=",
"index",
"[",
"(",
"-",
"1",
")",
"]",
"else",
":",
"det_limit_index",
"=",
"0",
"return",
"det_limit_index"
] | locates the corresponding detection limit for each observation . | train | false |
35,771 | def memcache_lock(prefix, expires=(60 * 60)):
def decorator(func):
@functools.wraps(func)
def wrapper(self, *args, **kwargs):
name = '_'.join(((prefix, func.__name__) + args))
lock = MemcacheLock(name, expires=expires)
if lock.locked():
log.warning(('Lock %s locked; ignoring call.' % name))
return
try:
lock.acquire()
except MemcacheLockException:
log.warning(('Aborting %s; lock acquisition failed.' % name))
return
else:
try:
return func(self, *args, **kwargs)
finally:
lock.release()
return wrapper
return decorator
| [
"def",
"memcache_lock",
"(",
"prefix",
",",
"expires",
"=",
"(",
"60",
"*",
"60",
")",
")",
":",
"def",
"decorator",
"(",
"func",
")",
":",
"@",
"functools",
".",
"wraps",
"(",
"func",
")",
"def",
"wrapper",
"(",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
":",
"name",
"=",
"'_'",
".",
"join",
"(",
"(",
"(",
"prefix",
",",
"func",
".",
"__name__",
")",
"+",
"args",
")",
")",
"lock",
"=",
"MemcacheLock",
"(",
"name",
",",
"expires",
"=",
"expires",
")",
"if",
"lock",
".",
"locked",
"(",
")",
":",
"log",
".",
"warning",
"(",
"(",
"'Lock %s locked; ignoring call.'",
"%",
"name",
")",
")",
"return",
"try",
":",
"lock",
".",
"acquire",
"(",
")",
"except",
"MemcacheLockException",
":",
"log",
".",
"warning",
"(",
"(",
"'Aborting %s; lock acquisition failed.'",
"%",
"name",
")",
")",
"return",
"else",
":",
"try",
":",
"return",
"func",
"(",
"self",
",",
"*",
"args",
",",
"**",
"kwargs",
")",
"finally",
":",
"lock",
".",
"release",
"(",
")",
"return",
"wrapper",
"return",
"decorator"
] | decorator that only allows one instance of the same command to run at a time . | train | false |
35,772 | def formstyle_inline(form, fields):
if (len(fields) != 2):
raise RuntimeError('Not possible')
(id, label, controls, help) = fields[0]
submit_button = fields[1][2]
return CAT(DIV(controls, _style='display:inline'), submit_button)
| [
"def",
"formstyle_inline",
"(",
"form",
",",
"fields",
")",
":",
"if",
"(",
"len",
"(",
"fields",
")",
"!=",
"2",
")",
":",
"raise",
"RuntimeError",
"(",
"'Not possible'",
")",
"(",
"id",
",",
"label",
",",
"controls",
",",
"help",
")",
"=",
"fields",
"[",
"0",
"]",
"submit_button",
"=",
"fields",
"[",
"1",
"]",
"[",
"2",
"]",
"return",
"CAT",
"(",
"DIV",
"(",
"controls",
",",
"_style",
"=",
"'display:inline'",
")",
",",
"submit_button",
")"
] | divs only . | train | false |
35,773 | @handle_response_format
@treeio_login_required
def document_edit(request, document_id, response_format='html'):
document = get_object_or_404(Document, pk=document_id)
if (not request.user.profile.has_permission(document, mode='w')):
return user_denied(request, message="You don't have access to this Document")
if request.POST:
if ('cancel' not in request.POST):
form = DocumentForm(request.user.profile, None, request.POST, instance=document)
if form.is_valid():
document = form.save()
return HttpResponseRedirect(reverse('documents_document_view', args=[document.id]))
else:
return HttpResponseRedirect(reverse('documents_document_view', args=[document.id]))
else:
form = DocumentForm(request.user.profile, None, instance=document)
context = _get_default_context(request)
context.update({'form': form, 'document': document})
return render_to_response('documents/document_edit', context, context_instance=RequestContext(request), response_format=response_format)
| [
"@",
"handle_response_format",
"@",
"treeio_login_required",
"def",
"document_edit",
"(",
"request",
",",
"document_id",
",",
"response_format",
"=",
"'html'",
")",
":",
"document",
"=",
"get_object_or_404",
"(",
"Document",
",",
"pk",
"=",
"document_id",
")",
"if",
"(",
"not",
"request",
".",
"user",
".",
"profile",
".",
"has_permission",
"(",
"document",
",",
"mode",
"=",
"'w'",
")",
")",
":",
"return",
"user_denied",
"(",
"request",
",",
"message",
"=",
"\"You don't have access to this Document\"",
")",
"if",
"request",
".",
"POST",
":",
"if",
"(",
"'cancel'",
"not",
"in",
"request",
".",
"POST",
")",
":",
"form",
"=",
"DocumentForm",
"(",
"request",
".",
"user",
".",
"profile",
",",
"None",
",",
"request",
".",
"POST",
",",
"instance",
"=",
"document",
")",
"if",
"form",
".",
"is_valid",
"(",
")",
":",
"document",
"=",
"form",
".",
"save",
"(",
")",
"return",
"HttpResponseRedirect",
"(",
"reverse",
"(",
"'documents_document_view'",
",",
"args",
"=",
"[",
"document",
".",
"id",
"]",
")",
")",
"else",
":",
"return",
"HttpResponseRedirect",
"(",
"reverse",
"(",
"'documents_document_view'",
",",
"args",
"=",
"[",
"document",
".",
"id",
"]",
")",
")",
"else",
":",
"form",
"=",
"DocumentForm",
"(",
"request",
".",
"user",
".",
"profile",
",",
"None",
",",
"instance",
"=",
"document",
")",
"context",
"=",
"_get_default_context",
"(",
"request",
")",
"context",
".",
"update",
"(",
"{",
"'form'",
":",
"form",
",",
"'document'",
":",
"document",
"}",
")",
"return",
"render_to_response",
"(",
"'documents/document_edit'",
",",
"context",
",",
"context_instance",
"=",
"RequestContext",
"(",
"request",
")",
",",
"response_format",
"=",
"response_format",
")"
] | document edit page . | train | false |
35,774 | def key_pair_count_by_user(context, user_id):
return IMPL.key_pair_count_by_user(context, user_id)
| [
"def",
"key_pair_count_by_user",
"(",
"context",
",",
"user_id",
")",
":",
"return",
"IMPL",
".",
"key_pair_count_by_user",
"(",
"context",
",",
"user_id",
")"
] | count number of key pairs for the given user id . | train | false |
35,777 | def renew_by(name, window=None):
return _renew_by(name, window).isoformat()
| [
"def",
"renew_by",
"(",
"name",
",",
"window",
"=",
"None",
")",
":",
"return",
"_renew_by",
"(",
"name",
",",
"window",
")",
".",
"isoformat",
"(",
")"
] | date in iso format when a certificate should first be renewed . | train | false |
35,778 | def http_date(timestamp=None):
return _dump_date(timestamp, ' ')
| [
"def",
"http_date",
"(",
"timestamp",
"=",
"None",
")",
":",
"return",
"_dump_date",
"(",
"timestamp",
",",
"' '",
")"
] | formats the time to match the rfc1123 date format . | train | false |
35,780 | def libvlc_media_get_stats(p_md, p_stats):
f = (_Cfunctions.get('libvlc_media_get_stats', None) or _Cfunction('libvlc_media_get_stats', ((1,), (1,)), None, ctypes.c_int, Media, ctypes.POINTER(MediaStats)))
return f(p_md, p_stats)
| [
"def",
"libvlc_media_get_stats",
"(",
"p_md",
",",
"p_stats",
")",
":",
"f",
"=",
"(",
"_Cfunctions",
".",
"get",
"(",
"'libvlc_media_get_stats'",
",",
"None",
")",
"or",
"_Cfunction",
"(",
"'libvlc_media_get_stats'",
",",
"(",
"(",
"1",
",",
")",
",",
"(",
"1",
",",
")",
")",
",",
"None",
",",
"ctypes",
".",
"c_int",
",",
"Media",
",",
"ctypes",
".",
"POINTER",
"(",
"MediaStats",
")",
")",
")",
"return",
"f",
"(",
"p_md",
",",
"p_stats",
")"
] | get the current statistics about the media . | train | true |
35,781 | def _get_proxy_auth(proxy_settings):
proxy_username = None
proxy_password = None
proxy_username = os.environ.get('proxy-username')
if (not proxy_username):
proxy_username = os.environ.get('proxy_username')
proxy_password = os.environ.get('proxy-password')
if (not proxy_password):
proxy_password = os.environ.get('proxy_password')
if (not proxy_username):
if ('@' in proxy_settings):
protocol_and_proxy_auth = proxy_settings.split('@')[0].split(':')
if (len(protocol_and_proxy_auth) == 3):
proxy_username = protocol_and_proxy_auth[1].lstrip('/')
proxy_password = protocol_and_proxy_auth[2]
elif (len(protocol_and_proxy_auth) == 2):
proxy_username = protocol_and_proxy_auth[0]
proxy_password = protocol_and_proxy_auth[1]
if proxy_username:
user_auth = base64.encodestring(('%s:%s' % (proxy_username, proxy_password)))
return ('Basic %s\r\n' % user_auth.strip())
else:
return ''
| [
"def",
"_get_proxy_auth",
"(",
"proxy_settings",
")",
":",
"proxy_username",
"=",
"None",
"proxy_password",
"=",
"None",
"proxy_username",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'proxy-username'",
")",
"if",
"(",
"not",
"proxy_username",
")",
":",
"proxy_username",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'proxy_username'",
")",
"proxy_password",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'proxy-password'",
")",
"if",
"(",
"not",
"proxy_password",
")",
":",
"proxy_password",
"=",
"os",
".",
"environ",
".",
"get",
"(",
"'proxy_password'",
")",
"if",
"(",
"not",
"proxy_username",
")",
":",
"if",
"(",
"'@'",
"in",
"proxy_settings",
")",
":",
"protocol_and_proxy_auth",
"=",
"proxy_settings",
".",
"split",
"(",
"'@'",
")",
"[",
"0",
"]",
".",
"split",
"(",
"':'",
")",
"if",
"(",
"len",
"(",
"protocol_and_proxy_auth",
")",
"==",
"3",
")",
":",
"proxy_username",
"=",
"protocol_and_proxy_auth",
"[",
"1",
"]",
".",
"lstrip",
"(",
"'/'",
")",
"proxy_password",
"=",
"protocol_and_proxy_auth",
"[",
"2",
"]",
"elif",
"(",
"len",
"(",
"protocol_and_proxy_auth",
")",
"==",
"2",
")",
":",
"proxy_username",
"=",
"protocol_and_proxy_auth",
"[",
"0",
"]",
"proxy_password",
"=",
"protocol_and_proxy_auth",
"[",
"1",
"]",
"if",
"proxy_username",
":",
"user_auth",
"=",
"base64",
".",
"encodestring",
"(",
"(",
"'%s:%s'",
"%",
"(",
"proxy_username",
",",
"proxy_password",
")",
")",
")",
"return",
"(",
"'Basic %s\\r\\n'",
"%",
"user_auth",
".",
"strip",
"(",
")",
")",
"else",
":",
"return",
"''"
] | returns proxy authentication string for header . | train | false |
35,784 | def make_params(**kwargs):
return {k: v for (k, v) in kwargs.items() if (v is not None)}
| [
"def",
"make_params",
"(",
"**",
"kwargs",
")",
":",
"return",
"{",
"k",
":",
"v",
"for",
"(",
"k",
",",
"v",
")",
"in",
"kwargs",
".",
"items",
"(",
")",
"if",
"(",
"v",
"is",
"not",
"None",
")",
"}"
] | helper to create a params dict . | train | false |
35,785 | def test_iht_fit_sample_wrong_class_obj():
from sklearn.cluster import KMeans
est = KMeans()
iht = InstanceHardnessThreshold(estimator=est, random_state=RND_SEED)
assert_raises(ValueError, iht.fit_sample, X, Y)
| [
"def",
"test_iht_fit_sample_wrong_class_obj",
"(",
")",
":",
"from",
"sklearn",
".",
"cluster",
"import",
"KMeans",
"est",
"=",
"KMeans",
"(",
")",
"iht",
"=",
"InstanceHardnessThreshold",
"(",
"estimator",
"=",
"est",
",",
"random_state",
"=",
"RND_SEED",
")",
"assert_raises",
"(",
"ValueError",
",",
"iht",
".",
"fit_sample",
",",
"X",
",",
"Y",
")"
] | test either if an error is raised while passing a wrong classifier object . | train | false |
35,787 | def _datetime_to_json(value):
if isinstance(value, datetime.datetime):
value = _datetime_to_rfc3339(value)
return value
| [
"def",
"_datetime_to_json",
"(",
"value",
")",
":",
"if",
"isinstance",
"(",
"value",
",",
"datetime",
".",
"datetime",
")",
":",
"value",
"=",
"_datetime_to_rfc3339",
"(",
"value",
")",
"return",
"value"
] | coerce value to an json-compatible representation . | train | false |
35,788 | def is_ccx_course(course_key):
return (isinstance(course_key, CCXLocator) or isinstance(course_key, CCXBlockUsageLocator))
| [
"def",
"is_ccx_course",
"(",
"course_key",
")",
":",
"return",
"(",
"isinstance",
"(",
"course_key",
",",
"CCXLocator",
")",
"or",
"isinstance",
"(",
"course_key",
",",
"CCXBlockUsageLocator",
")",
")"
] | check whether the course locator maps to a ccx course; this is important because we dont allow access to ccx courses in studio . | train | false |
35,790 | def password_data(username, password):
return {'username': username, 'password': hashlib.sha1(password.encode('utf-8')).hexdigest(), 'passwordMd5': hashlib.md5(password.encode('utf-8')).hexdigest()}
| [
"def",
"password_data",
"(",
"username",
",",
"password",
")",
":",
"return",
"{",
"'username'",
":",
"username",
",",
"'password'",
":",
"hashlib",
".",
"sha1",
"(",
"password",
".",
"encode",
"(",
"'utf-8'",
")",
")",
".",
"hexdigest",
"(",
")",
",",
"'passwordMd5'",
":",
"hashlib",
".",
"md5",
"(",
"password",
".",
"encode",
"(",
"'utf-8'",
")",
")",
".",
"hexdigest",
"(",
")",
"}"
] | returns a dict with username and its encoded password . | train | false |
35,791 | def get_navigator_server_url():
return get_conf().get(_CONF_NAVIGATOR_SERVER_URL, 'http://localhost:7187')
| [
"def",
"get_navigator_server_url",
"(",
")",
":",
"return",
"get_conf",
"(",
")",
".",
"get",
"(",
"_CONF_NAVIGATOR_SERVER_URL",
",",
"'http://localhost:7187'",
")"
] | returns the navigator . | train | false |
35,792 | def _merge_flags(new_flags, old_flags=None, conf='any'):
if (not old_flags):
old_flags = []
args = [old_flags, new_flags]
if (conf == 'accept_keywords'):
tmp = (new_flags + [i for i in old_flags if _check_accept_keywords(new_flags, i)])
else:
tmp = portage.flatten(args)
flags = {}
for flag in tmp:
if (flag[0] == '-'):
flags[flag[1:]] = False
else:
flags[flag] = True
tmp = []
for (key, val) in six.iteritems(flags):
if val:
tmp.append(key)
else:
tmp.append(('-' + key))
tmp.sort(cmp=(lambda x, y: cmp(x.lstrip('-'), y.lstrip('-'))))
return tmp
| [
"def",
"_merge_flags",
"(",
"new_flags",
",",
"old_flags",
"=",
"None",
",",
"conf",
"=",
"'any'",
")",
":",
"if",
"(",
"not",
"old_flags",
")",
":",
"old_flags",
"=",
"[",
"]",
"args",
"=",
"[",
"old_flags",
",",
"new_flags",
"]",
"if",
"(",
"conf",
"==",
"'accept_keywords'",
")",
":",
"tmp",
"=",
"(",
"new_flags",
"+",
"[",
"i",
"for",
"i",
"in",
"old_flags",
"if",
"_check_accept_keywords",
"(",
"new_flags",
",",
"i",
")",
"]",
")",
"else",
":",
"tmp",
"=",
"portage",
".",
"flatten",
"(",
"args",
")",
"flags",
"=",
"{",
"}",
"for",
"flag",
"in",
"tmp",
":",
"if",
"(",
"flag",
"[",
"0",
"]",
"==",
"'-'",
")",
":",
"flags",
"[",
"flag",
"[",
"1",
":",
"]",
"]",
"=",
"False",
"else",
":",
"flags",
"[",
"flag",
"]",
"=",
"True",
"tmp",
"=",
"[",
"]",
"for",
"(",
"key",
",",
"val",
")",
"in",
"six",
".",
"iteritems",
"(",
"flags",
")",
":",
"if",
"val",
":",
"tmp",
".",
"append",
"(",
"key",
")",
"else",
":",
"tmp",
".",
"append",
"(",
"(",
"'-'",
"+",
"key",
")",
")",
"tmp",
".",
"sort",
"(",
"cmp",
"=",
"(",
"lambda",
"x",
",",
"y",
":",
"cmp",
"(",
"x",
".",
"lstrip",
"(",
"'-'",
")",
",",
"y",
".",
"lstrip",
"(",
"'-'",
")",
")",
")",
")",
"return",
"tmp"
] | merges multiple lists of flags removing duplicates and resolving conflicts giving priority to lasts lists . | train | true |
35,793 | def pmonitorTest(N=3, seconds=10):
topo = SingleSwitchTopo(N)
net = Mininet(topo)
net.start()
hosts = net.hosts
info('Starting test...\n')
server = hosts[0]
popens = {}
for h in hosts:
popens[h] = h.popen('ping', server.IP())
info('Monitoring output for', seconds, 'seconds\n')
endTime = (time() + seconds)
for (h, line) in pmonitor(popens, timeoutms=500):
if h:
info(('<%s>: %s' % (h.name, line)))
if (time() >= endTime):
for p in popens.values():
p.send_signal(SIGINT)
net.stop()
| [
"def",
"pmonitorTest",
"(",
"N",
"=",
"3",
",",
"seconds",
"=",
"10",
")",
":",
"topo",
"=",
"SingleSwitchTopo",
"(",
"N",
")",
"net",
"=",
"Mininet",
"(",
"topo",
")",
"net",
".",
"start",
"(",
")",
"hosts",
"=",
"net",
".",
"hosts",
"info",
"(",
"'Starting test...\\n'",
")",
"server",
"=",
"hosts",
"[",
"0",
"]",
"popens",
"=",
"{",
"}",
"for",
"h",
"in",
"hosts",
":",
"popens",
"[",
"h",
"]",
"=",
"h",
".",
"popen",
"(",
"'ping'",
",",
"server",
".",
"IP",
"(",
")",
")",
"info",
"(",
"'Monitoring output for'",
",",
"seconds",
",",
"'seconds\\n'",
")",
"endTime",
"=",
"(",
"time",
"(",
")",
"+",
"seconds",
")",
"for",
"(",
"h",
",",
"line",
")",
"in",
"pmonitor",
"(",
"popens",
",",
"timeoutms",
"=",
"500",
")",
":",
"if",
"h",
":",
"info",
"(",
"(",
"'<%s>: %s'",
"%",
"(",
"h",
".",
"name",
",",
"line",
")",
")",
")",
"if",
"(",
"time",
"(",
")",
">=",
"endTime",
")",
":",
"for",
"p",
"in",
"popens",
".",
"values",
"(",
")",
":",
"p",
".",
"send_signal",
"(",
"SIGINT",
")",
"net",
".",
"stop",
"(",
")"
] | run pings and monitor multiple hosts using pmonitor . | train | false |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.