source_code stringlengths 45 7.24k | function_name stringlengths 1 111 | messages list |
|---|---|---|
def x(self, template_name, variables=None):
if variables is None:
variables = {}
template = self._engine.get_template(template_name)
return template.render(**variables)
| render | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, template_name, variables=None):\n if variables is None:\n variables = {}\n template = self._engine.get_template(template_name)\n return template.render(**variables)\n",
"role": "user"
},
{
"content": "render",
"role": "assistant"
}
] |
def x(self, source, variables=None):
if variables is None:
variables = {}
template = self._engine.from_string(source)
return template.render(**variables)
| render_source | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, source, variables=None):\n if variables is None:\n variables = {}\n template = self._engine.from_string(source)\n return template.render(**variables)\n",
"role": "user"
},
{
"content": "render_source",
"role": "assistant"
}
] |
def x(url_template, match_whole_str=False, converters=None,
default_converter='string', anonymous=False):
builder_params = []
url_params = {}
result = '^'
parts = _split_pattern.split(url_template)
for i, part in enumerate(parts):
is_url_pattern = _static_url_pattern.match(part)
if is_url_pattern:
result += re.escape(urlquote(part))
builder_params.append(part)
continue
is_converter = _converter_pattern.match(part)
if is_converter:
groups = is_converter.groupdict()
converter_name = groups['converter'] or default_converter
conv_object = init_converter(converters[converter_name], groups
['args'])
variable = groups['variable']
builder_params.append((variable, conv_object))
url_params[variable] = conv_object
if anonymous:
result += conv_object.regex
else:
result += '(?P<{}>{})'.format(variable, conv_object.regex)
continue
raise ValueError('Incorrect url template {!r}'.format(url_template))
if match_whole_str:
result += '$'
return re.compile(result), url_params, builder_params
| construct_re | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url_template, match_whole_str=False, converters=None,\n default_converter='string', anonymous=False):\n builder_params = []\n url_params = {}\n result = '^'\n parts = _split_pattern.split(url_template)\n for i, part in enumerate(parts):\n is_url_pattern = _static_url_pattern.match(part)\n if is_url_pattern:\n result += re.escape(urlquote(part))\n builder_params.append(part)\n continue\n is_converter = _converter_pattern.match(part)\n if is_converter:\n groups = is_converter.groupdict()\n converter_name = groups['converter'] or default_converter\n conv_object = init_converter(converters[converter_name], groups\n ['args'])\n variable = groups['variable']\n builder_params.append((variable, conv_object))\n url_params[variable] = conv_object\n if anonymous:\n result += conv_object.regex\n else:\n result += '(?P<{}>{})'.format(variable, conv_object.regex)\n continue\n raise ValueError('Incorrect url template {!r}'.format(url_template))\n if match_whole_str:\n result += '$'\n return re.compile(result), url_params, builder_params\n",
"role": "user"
},
{
"content": "construct_re",
"role": "assistant"
}
] |
def x(self, path, **kw):
m = self._pattern.match(path)
if m:
kwargs = m.groupdict()
for url_arg_name, value_urlencoded in kwargs.items():
conv_obj = self._url_params[url_arg_name]
unicode_value = unquote(value_urlencoded)
if isinstance(unicode_value, six.binary_type):
unicode_value = unicode_value.decode('utf-8', 'replace')
try:
kwargs[url_arg_name] = conv_obj.to_python(unicode_value, **kw)
except ConvertError as err:
logger.debug('ConvertError in parameter "%s" by %r, value "%s"'
, url_arg_name, err.converter.__class__, err.value)
return None, {}
return m.group(), kwargs
return None, {}
| match | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path, **kw):\n m = self._pattern.match(path)\n if m:\n kwargs = m.groupdict()\n for url_arg_name, value_urlencoded in kwargs.items():\n conv_obj = self._url_params[url_arg_name]\n unicode_value = unquote(value_urlencoded)\n if isinstance(unicode_value, six.binary_type):\n unicode_value = unicode_value.decode('utf-8', 'replace')\n try:\n kwargs[url_arg_name] = conv_obj.to_python(unicode_value, **kw)\n except ConvertError as err:\n logger.debug('ConvertError in parameter \"%s\" by %r, value \"%s\"'\n , url_arg_name, err.converter.__class__, err.value)\n return None, {}\n return m.group(), kwargs\n return None, {}\n",
"role": "user"
},
{
"content": "match",
"role": "assistant"
}
] |
def x(bundle, force=False, force_restricted=False):
if not ckan:
raise EnvironmentError(MISSING_CREDENTIALS_MSG)
try:
ckan.action.package_create(**_convert_bundle(bundle))
except ckanapi.ValidationError:
if force:
logger.warning(
'{} dataset already exported, but new export forced. Continue to export dataset stuff.'
.format(bundle.dataset))
else:
raise
access = bundle.dataset.config.metadata.about.access
if access == 'restricted' and force_restricted:
access = 'private'
assert access, 'CKAN publishing requires access level.'
if access in ('internal', 'controlled', 'restricted', 'census'):
raise UnpublishedAccessError(
'{} dataset can not be published because of {} access.'.format(
bundle.dataset.vid, bundle.dataset.config.metadata.about.access))
elif access == 'public':
user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.
vid.lower(), 'roles': ['editor']}, {'user': 'logged_in',
'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}]
elif access == 'registered':
user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.
vid.lower(), 'roles': []}, {'user': 'logged_in',
'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}]
elif access in ('private', 'licensed', 'test'):
user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.
vid.lower(), 'roles': []}, {'user': 'logged_in',
'domain_object': bundle.dataset.vid.lower(), 'roles': []}]
organization_users = ckan.action.organization_show(id=CKAN_CONFIG.
organization)['users']
for user in organization_users:
user_roles.append({'user': user['id'], 'domain_object': bundle.
dataset.vid.lower(), 'roles': ['editor']}),
for role in user_roles:
ckan.action.user_role_update(**role)
for partition in bundle.partitions:
ckan.action.resource_create(**_convert_partition(partition))
ckan.action.resource_create(**_convert_schema(bundle))
for name, external in six.iteritems(bundle.dataset.config.metadata.
external_documentation):
ckan.action.resource_create(**_convert_external(bundle, name, external)
)
| export | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bundle, force=False, force_restricted=False):\n if not ckan:\n raise EnvironmentError(MISSING_CREDENTIALS_MSG)\n try:\n ckan.action.package_create(**_convert_bundle(bundle))\n except ckanapi.ValidationError:\n if force:\n logger.warning(\n '{} dataset already exported, but new export forced. Continue to export dataset stuff.'\n .format(bundle.dataset))\n else:\n raise\n access = bundle.dataset.config.metadata.about.access\n if access == 'restricted' and force_restricted:\n access = 'private'\n assert access, 'CKAN publishing requires access level.'\n if access in ('internal', 'controlled', 'restricted', 'census'):\n raise UnpublishedAccessError(\n '{} dataset can not be published because of {} access.'.format(\n bundle.dataset.vid, bundle.dataset.config.metadata.about.access))\n elif access == 'public':\n user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.\n vid.lower(), 'roles': ['editor']}, {'user': 'logged_in',\n 'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}]\n elif access == 'registered':\n user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.\n vid.lower(), 'roles': []}, {'user': 'logged_in',\n 'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}]\n elif access in ('private', 'licensed', 'test'):\n user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.\n vid.lower(), 'roles': []}, {'user': 'logged_in',\n 'domain_object': bundle.dataset.vid.lower(), 'roles': []}]\n organization_users = ckan.action.organization_show(id=CKAN_CONFIG.\n organization)['users']\n for user in organization_users:\n user_roles.append({'user': user['id'], 'domain_object': bundle.\n dataset.vid.lower(), 'roles': ['editor']}),\n for role in user_roles:\n ckan.action.user_role_update(**role)\n for partition in bundle.partitions:\n ckan.action.resource_create(**_convert_partition(partition))\n ckan.action.resource_create(**_convert_schema(bundle))\n for name, external in six.iteritems(bundle.dataset.config.metadata.\n external_documentation):\n ckan.action.resource_create(**_convert_external(bundle, name, external)\n )\n",
"role": "user"
},
{
"content": "export",
"role": "assistant"
}
] |
def x(bundle):
if not ckan:
raise EnvironmentError(MISSING_CREDENTIALS_MSG)
params = {'q': 'name:{}'.format(bundle.dataset.vid.lower())}
resp = ckan.action.package_search(**params)
return len(resp['results']) > 0
| is_exported | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bundle):\n if not ckan:\n raise EnvironmentError(MISSING_CREDENTIALS_MSG)\n params = {'q': 'name:{}'.format(bundle.dataset.vid.lower())}\n resp = ckan.action.package_search(**params)\n return len(resp['results']) > 0\n",
"role": "user"
},
{
"content": "is_exported",
"role": "assistant"
}
] |
def x(bundle):
meta = bundle.dataset.config.metadata
notes = ''
for f in bundle.dataset.files:
if f.path.endswith('documentation.md'):
contents = f.unpacked_contents
if isinstance(contents, six.binary_type):
contents = contents.decode('utf-8')
notes = json.dumps(contents)
break
ret = {'name': bundle.dataset.vid.lower(), 'title': meta.about.title,
'author': meta.contacts.wrangler.name, 'author_email': meta.
contacts.wrangler.email, 'maintainer': meta.contacts.maintainer.
name, 'maintainer_email': meta.contacts.maintainer.email,
'license_id': '', 'notes': notes, 'url': meta.identity.source,
'version': bundle.dataset.version, 'state': 'active', 'owner_org':
CKAN_CONFIG['organization']}
return ret
| convert_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bundle):\n meta = bundle.dataset.config.metadata\n notes = ''\n for f in bundle.dataset.files:\n if f.path.endswith('documentation.md'):\n contents = f.unpacked_contents\n if isinstance(contents, six.binary_type):\n contents = contents.decode('utf-8')\n notes = json.dumps(contents)\n break\n ret = {'name': bundle.dataset.vid.lower(), 'title': meta.about.title,\n 'author': meta.contacts.wrangler.name, 'author_email': meta.\n contacts.wrangler.email, 'maintainer': meta.contacts.maintainer.\n name, 'maintainer_email': meta.contacts.maintainer.email,\n 'license_id': '', 'notes': notes, 'url': meta.identity.source,\n 'version': bundle.dataset.version, 'state': 'active', 'owner_org':\n CKAN_CONFIG['organization']}\n return ret\n",
"role": "user"
},
{
"content": "convert_bundle",
"role": "assistant"
}
] |
def x(partition):
csvfile = six.StringIO()
writer = unicodecsv.writer(csvfile)
headers = partition.datafile.headers
if headers:
writer.writerow(headers)
for row in partition:
writer.writerow([row[h] for h in headers])
csvfile.seek(0)
ret = {'package_id': partition.dataset.vid.lower(), 'url':
'http://example.com', 'revision_id': '', 'description': partition.
description or '', 'format': 'text/csv', 'hash': '', 'name':
partition.name, 'resource_type': '', 'mimetype': 'text/csv',
'mimetype_inner': '', 'webstore_url': '', 'cache_url': '', 'upload':
csvfile}
return ret
| convert_partition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(partition):\n csvfile = six.StringIO()\n writer = unicodecsv.writer(csvfile)\n headers = partition.datafile.headers\n if headers:\n writer.writerow(headers)\n for row in partition:\n writer.writerow([row[h] for h in headers])\n csvfile.seek(0)\n ret = {'package_id': partition.dataset.vid.lower(), 'url':\n 'http://example.com', 'revision_id': '', 'description': partition.\n description or '', 'format': 'text/csv', 'hash': '', 'name':\n partition.name, 'resource_type': '', 'mimetype': 'text/csv',\n 'mimetype_inner': '', 'webstore_url': '', 'cache_url': '', 'upload':\n csvfile}\n return ret\n",
"role": "user"
},
{
"content": "convert_partition",
"role": "assistant"
}
] |
def x(bundle):
schema_csv = None
for f in bundle.dataset.files:
if f.path.endswith('schema.csv'):
contents = f.unpacked_contents
if isinstance(contents, six.binary_type):
contents = contents.decode('utf-8')
schema_csv = six.StringIO(contents)
schema_csv.seek(0)
break
ret = {'package_id': bundle.dataset.vid.lower(), 'url':
'http://example.com', 'revision_id': '', 'description':
'Schema of the dataset tables.', 'format': 'text/csv', 'hash': '',
'name': 'schema', 'upload': schema_csv}
return ret
| convert_schema | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bundle):\n schema_csv = None\n for f in bundle.dataset.files:\n if f.path.endswith('schema.csv'):\n contents = f.unpacked_contents\n if isinstance(contents, six.binary_type):\n contents = contents.decode('utf-8')\n schema_csv = six.StringIO(contents)\n schema_csv.seek(0)\n break\n ret = {'package_id': bundle.dataset.vid.lower(), 'url':\n 'http://example.com', 'revision_id': '', 'description':\n 'Schema of the dataset tables.', 'format': 'text/csv', 'hash': '',\n 'name': 'schema', 'upload': schema_csv}\n return ret\n",
"role": "user"
},
{
"content": "convert_schema",
"role": "assistant"
}
] |
def x(bundle, name, external):
ret = {'package_id': bundle.dataset.vid.lower(), 'url': external.url,
'description': external.description, 'name': name}
return ret
| convert_external | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bundle, name, external):\n ret = {'package_id': bundle.dataset.vid.lower(), 'url': external.url,\n 'description': external.description, 'name': name}\n return ret\n",
"role": "user"
},
{
"content": "convert_external",
"role": "assistant"
}
] |
def x(repository, revision, **kwargs):
if not kwargs.get('dependency_ids'):
kwargs['dependency_ids'] = []
if not kwargs.get('build_configuration_set_ids'):
kwargs['build_configuration_set_ids'] = []
if kwargs.get('generic_parameters'):
kwargs['generic_parameters'] = ast.literal_eval(kwargs.get(
'generic_parameters'))
if not kwargs.get('project'):
kwargs['project'] = pnc_api.projects.get_specific(kwargs.get(
'project_id')).content
if not kwargs.get('environment'):
kwargs['environment'] = pnc_api.environments.get_specific(kwargs.
get('build_environment_id')).content
build_configuration = create_build_conf_object(scm_revision=revision,
**kwargs)
repo_creation = swagger_client.RepositoryCreationUrlAutoRest()
repo_creation.scm_url = repository
repo_creation.build_configuration_rest = build_configuration
response = utils.checked_api_call(pnc_api.bpm,
'start_r_creation_task_with_single_url', body=repo_creation)
if response:
return response
| create_build_configuration_process | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(repository, revision, **kwargs):\n if not kwargs.get('dependency_ids'):\n kwargs['dependency_ids'] = []\n if not kwargs.get('build_configuration_set_ids'):\n kwargs['build_configuration_set_ids'] = []\n if kwargs.get('generic_parameters'):\n kwargs['generic_parameters'] = ast.literal_eval(kwargs.get(\n 'generic_parameters'))\n if not kwargs.get('project'):\n kwargs['project'] = pnc_api.projects.get_specific(kwargs.get(\n 'project_id')).content\n if not kwargs.get('environment'):\n kwargs['environment'] = pnc_api.environments.get_specific(kwargs.\n get('build_environment_id')).content\n build_configuration = create_build_conf_object(scm_revision=revision,\n **kwargs)\n repo_creation = swagger_client.RepositoryCreationUrlAutoRest()\n repo_creation.scm_url = repository\n repo_creation.build_configuration_rest = build_configuration\n response = utils.checked_api_call(pnc_api.bpm,\n 'start_r_creation_task_with_single_url', body=repo_creation)\n if response:\n return response\n",
"role": "user"
},
{
"content": "create_build_configuration_process",
"role": "assistant"
}
] |
def x(repository, no_sync=False):
repo = create_repository_configuration_raw(repository, no_sync)
if repo:
return utils.format_json(repo)
| create_repository_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(repository, no_sync=False):\n repo = create_repository_configuration_raw(repository, no_sync)\n if repo:\n return utils.format_json(repo)\n",
"role": "user"
},
{
"content": "create_repository_configuration",
"role": "assistant"
}
] |
def x(task_id, retry=30):
success_event_types = 'RC_CREATION_SUCCESS',
error_event_types = ('RC_REPO_CREATION_ERROR', 'RC_REPO_CLONE_ERROR',
'RC_CREATION_ERROR')
while retry > 0:
bpm_task = get_bpm_task_by_id(task_id)
if contains_event_type(bpm_task.content.events, success_event_types):
break
if contains_event_type(bpm_task.content.events, error_event_types):
logging.error('Creation of Repository Configuration failed')
logging.error(bpm_task.content)
return False
logging.info(
'Waiting until Repository Configuration creation task ' + str(
task_id) + ' finishes.')
time.sleep(10)
retry -= 1
return retry > 0
| wait_for_repo_creation | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(task_id, retry=30):\n success_event_types = 'RC_CREATION_SUCCESS',\n error_event_types = ('RC_REPO_CREATION_ERROR', 'RC_REPO_CLONE_ERROR',\n 'RC_CREATION_ERROR')\n while retry > 0:\n bpm_task = get_bpm_task_by_id(task_id)\n if contains_event_type(bpm_task.content.events, success_event_types):\n break\n if contains_event_type(bpm_task.content.events, error_event_types):\n logging.error('Creation of Repository Configuration failed')\n logging.error(bpm_task.content)\n return False\n logging.info(\n 'Waiting until Repository Configuration creation task ' + str(\n task_id) + ' finishes.')\n time.sleep(10)\n retry -= 1\n return retry > 0\n",
"role": "user"
},
{
"content": "wait_for_repo_creation",
"role": "assistant"
}
] |
def x(self):
args = self.format_args
if args is None or isinstance(args, dict) and self.count_field not in args:
raise TypeError('count is required')
return args[self.count_field] if isinstance(args, dict) else args
| count | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n args = self.format_args\n if args is None or isinstance(args, dict) and self.count_field not in args:\n raise TypeError('count is required')\n return args[self.count_field] if isinstance(args, dict) else args\n",
"role": "user"
},
{
"content": "count",
"role": "assistant"
}
] |
def x(milestone_id=None, output=None):
if not is_input_valid(milestone_id, output):
logging.error('invalid input')
return 1
create_work_dir(output)
download_sources_artifacts(milestone_id, output)
create_zip(output)
| generate_sources_zip | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(milestone_id=None, output=None):\n if not is_input_valid(milestone_id, output):\n logging.error('invalid input')\n return 1\n create_work_dir(output)\n download_sources_artifacts(milestone_id, output)\n create_zip(output)\n",
"role": "user"
},
{
"content": "generate_sources_zip",
"role": "assistant"
}
] |
def x(id):
response = utils.checked_api_call(pnc_api.repositories, 'get_specific',
id=id)
if response:
return response.content
| get_repository_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(id):\n response = utils.checked_api_call(pnc_api.repositories, 'get_specific',\n id=id)\n if response:\n return response.content\n",
"role": "user"
},
{
"content": "get_repository_configuration",
"role": "assistant"
}
] |
def x(id, external_repository=None, prebuild_sync=None):
to_update_id = id
rc_to_update = pnc_api.repositories.get_specific(id=to_update_id).content
if external_repository is None:
external_repository = rc_to_update.external_url
else:
rc_to_update.external_url = external_repository
if prebuild_sync is not None:
rc_to_update.pre_build_sync_enabled = prebuild_sync
if not external_repository and prebuild_sync:
logging.error(
'You cannot enable prebuild sync without external repository')
return
response = utils.checked_api_call(pnc_api.repositories, 'update', id=
to_update_id, body=rc_to_update)
if response:
return response.content
| update_repository_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(id, external_repository=None, prebuild_sync=None):\n to_update_id = id\n rc_to_update = pnc_api.repositories.get_specific(id=to_update_id).content\n if external_repository is None:\n external_repository = rc_to_update.external_url\n else:\n rc_to_update.external_url = external_repository\n if prebuild_sync is not None:\n rc_to_update.pre_build_sync_enabled = prebuild_sync\n if not external_repository and prebuild_sync:\n logging.error(\n 'You cannot enable prebuild sync without external repository')\n return\n response = utils.checked_api_call(pnc_api.repositories, 'update', id=\n to_update_id, body=rc_to_update)\n if response:\n return response.content\n",
"role": "user"
},
{
"content": "update_repository_configuration",
"role": "assistant"
}
] |
def x(page_size=200, page_index=0, sort='', q=''):
response = utils.checked_api_call(pnc_api.repositories, 'get_all',
page_size=page_size, page_index=page_index, sort=sort, q=q)
if response:
return utils.format_json_list(response.content)
| list_repository_configurations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(page_size=200, page_index=0, sort='', q=''):\n response = utils.checked_api_call(pnc_api.repositories, 'get_all',\n page_size=page_size, page_index=page_index, sort=sort, q=q)\n if response:\n return utils.format_json_list(response.content)\n",
"role": "user"
},
{
"content": "list_repository_configurations",
"role": "assistant"
}
] |
def x(url, page_size=10, page_index=0, sort=''):
content = search_repository_configuration_raw(url, page_size,
page_index, sort)
if content:
return utils.format_json_list(content)
| search_repository_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, page_size=10, page_index=0, sort=''):\n content = search_repository_configuration_raw(url, page_size,\n page_index, sort)\n if content:\n return utils.format_json_list(content)\n",
"role": "user"
},
{
"content": "search_repository_configuration",
"role": "assistant"
}
] |
def x(url, page_size=10, page_index=0, sort=''):
response = utils.checked_api_call(pnc_api.repositories, 'search',
page_size=page_size, page_index=page_index, sort=sort, search=url)
if response:
return response.content
| search_repository_configuration_raw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, page_size=10, page_index=0, sort=''):\n response = utils.checked_api_call(pnc_api.repositories, 'search',\n page_size=page_size, page_index=page_index, sort=sort, search=url)\n if response:\n return response.content\n",
"role": "user"
},
{
"content": "search_repository_configuration_raw",
"role": "assistant"
}
] |
def x(url, page_size=10, page_index=0, sort=''):
content = match_repository_configuration_raw(url, page_size, page_index,
sort)
if content:
return utils.format_json_list(content)
| match_repository_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(url, page_size=10, page_index=0, sort=''):\n content = match_repository_configuration_raw(url, page_size, page_index,\n sort)\n if content:\n return utils.format_json_list(content)\n",
"role": "user"
},
{
"content": "match_repository_configuration",
"role": "assistant"
}
] |
def x(self):
return self.env.template.render(self.template, form=self)
| render | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return self.env.template.render(self.template, form=self)\n",
"role": "user"
},
{
"content": "render",
"role": "assistant"
}
] |
def x(self, data):
self.raw_data = MultiDict(data)
self.errors = {}
for field in self.fields:
if field.writable:
self.python_data.update(field.accept())
else:
for name in field.field_names:
subfield = self.get_field(name)
value = self.python_data[subfield.name]
subfield.set_raw_value(self.raw_data, subfield.from_python(
value))
return self.is_valid
| accept | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, data):\n self.raw_data = MultiDict(data)\n self.errors = {}\n for field in self.fields:\n if field.writable:\n self.python_data.update(field.accept())\n else:\n for name in field.field_names:\n subfield = self.get_field(name)\n value = self.python_data[subfield.name]\n subfield.set_raw_value(self.raw_data, subfield.from_python(\n value))\n return self.is_valid\n",
"role": "user"
},
{
"content": "accept",
"role": "assistant"
}
] |
def x(self, compact=True):
data = MultiDict()
for field in self.fields:
raw_value = field.from_python(self.python_data[field.name])
field.set_raw_value(data, raw_value)
if compact:
data = MultiDict([(k, v) for k, v in data.items() if v])
return data
| get_data | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, compact=True):\n data = MultiDict()\n for field in self.fields:\n raw_value = field.from_python(self.python_data[field.name])\n field.set_raw_value(data, raw_value)\n if compact:\n data = MultiDict([(k, v) for k, v in data.items() if v])\n return data\n",
"role": "user"
},
{
"content": "get_data",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.add_configuration_with_http_info(id, **kwargs)
else:
data = self.add_configuration_with_http_info(id, **kwargs)
return data
| add_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.add_configuration_with_http_info(id, **kwargs)\n else:\n data = self.add_configuration_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "add_configuration",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.build_with_http_info(id, **kwargs)
else:
data = self.build_with_http_info(id, **kwargs)
return data
| build | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.build_with_http_info(id, **kwargs)\n else:\n data = self.build_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "build",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.build_versioned_with_http_info(id, **kwargs)
else:
data = self.build_versioned_with_http_info(id, **kwargs)
return data
| build_versioned | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.build_versioned_with_http_info(id, **kwargs)\n else:\n data = self.build_versioned_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "build_versioned",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.delete_specific_with_http_info(id, **kwargs)
else:
data = self.delete_specific_with_http_info(id, **kwargs)
return data
| delete_specific | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.delete_specific_with_http_info(id, **kwargs)\n else:\n data = self.delete_specific_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "delete_specific",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.get_all_build_config_set_records_with_http_info(id, **
kwargs)
else:
data = self.get_all_build_config_set_records_with_http_info(id, **
kwargs)
return data
| get_all_build_config_set_records | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_all_build_config_set_records_with_http_info(id, **\n kwargs)\n else:\n data = self.get_all_build_config_set_records_with_http_info(id, **\n kwargs)\n return data\n",
"role": "user"
},
{
"content": "get_all_build_config_set_records",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.get_build_records_with_http_info(id, **kwargs)
else:
data = self.get_build_records_with_http_info(id, **kwargs)
return data
| get_build_records | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_build_records_with_http_info(id, **kwargs)\n else:\n data = self.get_build_records_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "get_build_records",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.get_configurations_with_http_info(id, **kwargs)
else:
data = self.get_configurations_with_http_info(id, **kwargs)
return data
| get_configurations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_configurations_with_http_info(id, **kwargs)\n else:\n data = self.get_configurations_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "get_configurations",
"role": "assistant"
}
] |
def x(self, id, config_id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.remove_configuration_with_http_info(id, config_id, **kwargs
)
else:
data = self.remove_configuration_with_http_info(id, config_id, **kwargs
)
return data
| remove_configuration | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, config_id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.remove_configuration_with_http_info(id, config_id, **kwargs\n )\n else:\n data = self.remove_configuration_with_http_info(id, config_id, **kwargs\n )\n return data\n",
"role": "user"
},
{
"content": "remove_configuration",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.update_configurations_with_http_info(id, **kwargs)
else:
data = self.update_configurations_with_http_info(id, **kwargs)
return data
| update_configurations | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.update_configurations_with_http_info(id, **kwargs)\n else:\n data = self.update_configurations_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "update_configurations",
"role": "assistant"
}
] |
def x(name):
package = name.split('.')
return [os.path.abspath(os.path.join(x, *package)) for x in sys.path if
not os.path.exists(os.path.join(x, *(package + ['__init__.py'])))]
| pluginPackagePaths | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(name):\n package = name.split('.')\n return [os.path.abspath(os.path.join(x, *package)) for x in sys.path if\n not os.path.exists(os.path.join(x, *(package + ['__init__.py'])))]\n",
"role": "user"
},
{
"content": "pluginPackagePaths",
"role": "assistant"
}
] |
def x(func):
def x(self, *args, **kwargs):
return func(self._root_storage, *args, **kwargs)
return wrap
| storage_method | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(func):\n\n def x(self, *args, **kwargs):\n return func(self._root_storage, *args, **kwargs)\n return wrap\n",
"role": "user"
},
{
"content": "storage_method",
"role": "assistant"
}
] |
def x(self, template_name, **kw):
return self.env.get_template(template_name).render(**kw)
| render | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, template_name, **kw):\n return self.env.get_template(template_name).render(**kw)\n",
"role": "user"
},
{
"content": "render",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.cancel_all_builds_in_group_with_http_info(id, **kwargs)
else:
data = self.cancel_all_builds_in_group_with_http_info(id, **kwargs)
return data
| cancel_all_builds_in_group | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.cancel_all_builds_in_group_with_http_info(id, **kwargs)\n else:\n data = self.cancel_all_builds_in_group_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "cancel_all_builds_in_group",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.get_all_for_bc_with_http_info(id, **kwargs)
else:
data = self.get_all_for_bc_with_http_info(id, **kwargs)
return data
| get_all_for_bc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_all_for_bc_with_http_info(id, **kwargs)\n else:\n data = self.get_all_for_bc_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "get_all_for_bc",
"role": "assistant"
}
] |
def x(self, id, **kwargs):
kwargs['_return_http_data_only'] = True
if kwargs.get('callback'):
return self.get_all_for_bc_set_record_with_http_info(id, **kwargs)
else:
data = self.get_all_for_bc_set_record_with_http_info(id, **kwargs)
return data
| get_all_for_bc_set_record | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_all_for_bc_set_record_with_http_info(id, **kwargs)\n else:\n data = self.get_all_for_bc_set_record_with_http_info(id, **kwargs)\n return data\n",
"role": "user"
},
{
"content": "get_all_for_bc_set_record",
"role": "assistant"
}
] |
def x(typename, mapping):
def x(self, d, *args, **kwargs):
self._original_dict = d
def x(path):
if not callable(path) and len(path) < 1:
raise ValueError(
'Dict paths should be iterables with at least one key or callable objects that take one argument.'
)
def x(self):
cur_dict = self._original_dict
if callable(path):
return path(cur_dict)
return dict_value_from_path(cur_dict, path)
return getter
prop_mapping = dict((k, property(getter_from_dict_path(v))) for k, v in
mapping.iteritems())
prop_mapping['__init__'] = init
return type(typename, tuple(), prop_mapping)
| dictmapper | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(typename, mapping):\n\n def x(self, d, *args, **kwargs):\n self._original_dict = d\n\n def x(path):\n if not callable(path) and len(path) < 1:\n raise ValueError(\n 'Dict paths should be iterables with at least one key or callable objects that take one argument.'\n )\n\n def x(self):\n cur_dict = self._original_dict\n if callable(path):\n return path(cur_dict)\n return dict_value_from_path(cur_dict, path)\n return getter\n prop_mapping = dict((k, property(getter_from_dict_path(v))) for k, v in\n mapping.iteritems())\n prop_mapping['__init__'] = init\n return type(typename, tuple(), prop_mapping)\n",
"role": "user"
},
{
"content": "dictmapper",
"role": "assistant"
}
] |
def x(self):
return dict(config=self._config, search=self._search, echo=self._echo,
read_only=self.read_only)
| ctor_args | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return dict(config=self._config, search=self._search, echo=self._echo,\n read_only=self.read_only)\n",
"role": "user"
},
{
"content": "ctor_args",
"role": "assistant"
}
] |
def x(self, force=False):
from ambry.library.config import LibraryConfigSyncProxy
lcsp = LibraryConfigSyncProxy(self)
lcsp.sync(force=force)
| sync_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, force=False):\n from ambry.library.config import LibraryConfigSyncProxy\n lcsp = LibraryConfigSyncProxy(self)\n lcsp.sync(force=force)\n",
"role": "user"
},
{
"content": "sync_config",
"role": "assistant"
}
] |
def x(self):
import signal
def x(sig, frame):
self.log('Trace signal received')
self.log(''.join(traceback.format_stack(frame)))
signal.signal(signal.SIGUSR2, debug_trace)
| init_debug | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n import signal\n\n def x(sig, frame):\n self.log('Trace signal received')\n self.log(''.join(traceback.format_stack(frame)))\n signal.signal(signal.SIGUSR2, debug_trace)\n",
"role": "user"
},
{
"content": "init_debug",
"role": "assistant"
}
] |
def x(self, ref):
if not isinstance(ref, ObjectNumber):
on = ObjectNumber.parse(ref)
else:
on = ref
ds_on = on.as_dataset
return ds_on
| resolve_object_number | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref):\n if not isinstance(ref, ObjectNumber):\n on = ObjectNumber.parse(ref)\n else:\n on = ref\n ds_on = on.as_dataset\n return ds_on\n",
"role": "user"
},
{
"content": "resolve_object_number",
"role": "assistant"
}
] |
def x(self, ref, load_all=False, exception=True):
return self.database.dataset(ref, load_all=load_all, exception=exception)
| dataset | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref, load_all=False, exception=True):\n return self.database.dataset(ref, load_all=load_all, exception=exception)\n",
"role": "user"
},
{
"content": "dataset",
"role": "assistant"
}
] |
def x(self, assignment_class=None, **kwargs):
if not ('id' in kwargs and bool(kwargs['id'])
) or assignment_class is not None:
kwargs['id'] = self.number(assignment_class)
ds = self._db.new_dataset(**kwargs)
self._db.commit()
b = self.bundle(ds.vid)
b.state = Bundle.STATES.NEW
b.set_last_access(Bundle.STATES.NEW)
b.set_file_system(source_url=self._fs.source(b.identity.source_path),
build_url=self._fs.build(b.identity.source_path))
bs_meta = b.build_source_files.file(File.BSFILE.META)
bs_meta.set_defaults()
bs_meta.record_to_objects()
bs_meta.objects_to_record()
b.commit()
self._db.commit()
return b
| new_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, assignment_class=None, **kwargs):\n if not ('id' in kwargs and bool(kwargs['id'])\n ) or assignment_class is not None:\n kwargs['id'] = self.number(assignment_class)\n ds = self._db.new_dataset(**kwargs)\n self._db.commit()\n b = self.bundle(ds.vid)\n b.state = Bundle.STATES.NEW\n b.set_last_access(Bundle.STATES.NEW)\n b.set_file_system(source_url=self._fs.source(b.identity.source_path),\n build_url=self._fs.build(b.identity.source_path))\n bs_meta = b.build_source_files.file(File.BSFILE.META)\n bs_meta.set_defaults()\n bs_meta.record_to_objects()\n bs_meta.objects_to_record()\n b.commit()\n self._db.commit()\n return b\n",
"role": "user"
},
{
"content": "new_bundle",
"role": "assistant"
}
] |
def x(self, config):
identity = Identity.from_dict(config['identity'])
ds = self._db.dataset(identity.vid, exception=False)
if not ds:
ds = self._db.new_dataset(**identity.dict)
b = Bundle(ds, self)
b.commit()
b.state = Bundle.STATES.NEW
b.set_last_access(Bundle.STATES.NEW)
return b
| new_from_bundle_config | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, config):\n identity = Identity.from_dict(config['identity'])\n ds = self._db.dataset(identity.vid, exception=False)\n if not ds:\n ds = self._db.new_dataset(**identity.dict)\n b = Bundle(ds, self)\n b.commit()\n b.state = Bundle.STATES.NEW\n b.set_last_access(Bundle.STATES.NEW)\n return b\n",
"role": "user"
},
{
"content": "new_from_bundle_config",
"role": "assistant"
}
] |
def x(self, ref, capture_exceptions=False):
from ..orm.exc import NotFoundError
if isinstance(ref, Dataset):
ds = ref
else:
try:
ds = self._db.dataset(ref)
except NotFoundError:
ds = None
if not ds:
try:
p = self.partition(ref)
ds = p._bundle.dataset
except NotFoundError:
ds = None
if not ds:
raise NotFoundError('Failed to find dataset for ref: {}'.format(ref))
b = Bundle(ds, self)
b.capture_exceptions = capture_exceptions
return b
| bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref, capture_exceptions=False):\n from ..orm.exc import NotFoundError\n if isinstance(ref, Dataset):\n ds = ref\n else:\n try:\n ds = self._db.dataset(ref)\n except NotFoundError:\n ds = None\n if not ds:\n try:\n p = self.partition(ref)\n ds = p._bundle.dataset\n except NotFoundError:\n ds = None\n if not ds:\n raise NotFoundError('Failed to find dataset for ref: {}'.format(ref))\n b = Bundle(ds, self)\n b.capture_exceptions = capture_exceptions\n return b\n",
"role": "user"
},
{
"content": "bundle",
"role": "assistant"
}
] |
def x(self, ref, localize=False):
if not ref:
raise NotFoundError('No partition for empty ref')
try:
on = ObjectNumber.parse(ref)
ds_on = on.as_dataset
ds = self._db.dataset(ds_on)
self.database.session.refresh(ds)
p = ds.partition(ref)
except NotObjectNumberError:
q = self.database.session.query(Partition).filter(or_(Partition.
name == str(ref), Partition.vname == str(ref))).order_by(Partition
.vid.desc())
p = q.first()
if not p:
raise NotFoundError("No partition for ref: '{}'".format(ref))
b = self.bundle(p.d_vid)
p = b.wrap_partition(p)
if localize:
p.localize()
return p
| partition | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref, localize=False):\n if not ref:\n raise NotFoundError('No partition for empty ref')\n try:\n on = ObjectNumber.parse(ref)\n ds_on = on.as_dataset\n ds = self._db.dataset(ds_on)\n self.database.session.refresh(ds)\n p = ds.partition(ref)\n except NotObjectNumberError:\n q = self.database.session.query(Partition).filter(or_(Partition.\n name == str(ref), Partition.vname == str(ref))).order_by(Partition\n .vid.desc())\n p = q.first()\n if not p:\n raise NotFoundError(\"No partition for ref: '{}'\".format(ref))\n b = self.bundle(p.d_vid)\n p = b.wrap_partition(p)\n if localize:\n p.localize()\n return p\n",
"role": "user"
},
{
"content": "partition",
"role": "assistant"
}
] |
def x(self, ref):
try:
obj_number = ObjectNumber.parse(ref)
ds_obj_number = obj_number.as_dataset
dataset = self._db.dataset(ds_obj_number)
table = dataset.table(ref)
except NotObjectNumberError:
q = self.database.session.query(Table).filter(Table.name == str(ref)
).order_by(Table.vid.desc())
table = q.first()
if not table:
raise NotFoundError("No table for ref: '{}'".format(ref))
return table
| table | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref):\n try:\n obj_number = ObjectNumber.parse(ref)\n ds_obj_number = obj_number.as_dataset\n dataset = self._db.dataset(ds_obj_number)\n table = dataset.table(ref)\n except NotObjectNumberError:\n q = self.database.session.query(Table).filter(Table.name == str(ref)\n ).order_by(Table.vid.desc())\n table = q.first()\n if not table:\n raise NotFoundError(\"No table for ref: '{}'\".format(ref))\n return table\n",
"role": "user"
},
{
"content": "table",
"role": "assistant"
}
] |
def x(self, bundle):
from six import string_types
if isinstance(bundle, string_types):
bundle = self.bundle(bundle)
self.database.remove_dataset(bundle.dataset)
| remove | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, bundle):\n from six import string_types\n if isinstance(bundle, string_types):\n bundle = self.bundle(bundle)\n self.database.remove_dataset(bundle.dataset)\n",
"role": "user"
},
{
"content": "remove",
"role": "assistant"
}
] |
def x(self, b):
on = b.identity.on
on.revision = on.revision + 1
try:
extant = self.bundle(str(on))
if extant:
raise ConflictError('Already have a bundle with vid: {}'.format
(str(on)))
except NotFoundError:
pass
d = b.dataset.dict
d['revision'] = on.revision
d['vid'] = str(on)
del d['name']
del d['vname']
del d['version']
del d['fqname']
del d['cache_key']
ds = self.database.new_dataset(**d)
nb = self.bundle(ds.vid)
nb.set_file_system(source_url=b.source_fs.getsyspath('/'))
nb.state = Bundle.STATES.NEW
nb.commit()
for f in b.dataset.files:
assert f.major_type == f.MAJOR_TYPE.BUILDSOURCE
nb.dataset.files.append(nb.dataset.bsfile(f.minor_type, f.path).
update(f))
nb.build_source_files.file(File.BSFILE.META).record_to_objects()
nb.build_source_files.file(File.BSFILE.META).objects_to_record()
ds.commit()
return nb
| duplicate | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, b):\n on = b.identity.on\n on.revision = on.revision + 1\n try:\n extant = self.bundle(str(on))\n if extant:\n raise ConflictError('Already have a bundle with vid: {}'.format\n (str(on)))\n except NotFoundError:\n pass\n d = b.dataset.dict\n d['revision'] = on.revision\n d['vid'] = str(on)\n del d['name']\n del d['vname']\n del d['version']\n del d['fqname']\n del d['cache_key']\n ds = self.database.new_dataset(**d)\n nb = self.bundle(ds.vid)\n nb.set_file_system(source_url=b.source_fs.getsyspath('/'))\n nb.state = Bundle.STATES.NEW\n nb.commit()\n for f in b.dataset.files:\n assert f.major_type == f.MAJOR_TYPE.BUILDSOURCE\n nb.dataset.files.append(nb.dataset.bsfile(f.minor_type, f.path).\n update(f))\n nb.build_source_files.file(File.BSFILE.META).record_to_objects()\n nb.build_source_files.file(File.BSFILE.META).objects_to_record()\n ds.commit()\n return nb\n",
"role": "user"
},
{
"content": "duplicate",
"role": "assistant"
}
] |
def x(self, db_path, replace=True, cb=None):
from ambry.orm.exc import NotFoundError
db = Database('sqlite:///{}'.format(db_path))
db.open()
if len(db.datasets) == 0:
raise NotFoundError('Did not get a dataset in the {} bundle'.format
(db_path))
ds = db.dataset(db.datasets[0].vid)
assert ds is not None
assert ds._database
try:
b = self.bundle(ds.vid)
self.logger.info(
"Removing old bundle before checking in new one of same number: '{}'"
.format(ds.vid))
self.remove(b)
except NotFoundError:
pass
try:
self.dataset(ds.vid)
except NotFoundError:
self.database.copy_dataset(ds, cb=cb)
b = self.bundle(ds.vid)
b.commit()
self.search.index_bundle(b)
return b
| checkin_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, db_path, replace=True, cb=None):\n from ambry.orm.exc import NotFoundError\n db = Database('sqlite:///{}'.format(db_path))\n db.open()\n if len(db.datasets) == 0:\n raise NotFoundError('Did not get a dataset in the {} bundle'.format\n (db_path))\n ds = db.dataset(db.datasets[0].vid)\n assert ds is not None\n assert ds._database\n try:\n b = self.bundle(ds.vid)\n self.logger.info(\n \"Removing old bundle before checking in new one of same number: '{}'\"\n .format(ds.vid))\n self.remove(b)\n except NotFoundError:\n pass\n try:\n self.dataset(ds.vid)\n except NotFoundError:\n self.database.copy_dataset(ds, cb=cb)\n b = self.bundle(ds.vid)\n b.commit()\n self.search.index_bundle(b)\n return b\n",
"role": "user"
},
{
"content": "checkin_bundle",
"role": "assistant"
}
] |
def x(self, b, no_partitions=False):
raise DeprecationWarning("Don't use any more?")
from ambry.bundle.process import call_interval
remote_name = self.resolve_remote(b)
remote = self.remote(remote_name)
db_path = b.package()
with b.progress.start('checkin', 0, message='Check in bundle') as ps:
ps.add(message='Checking in bundle {} to {}'.format(b.identity.
vname, remote))
db_ck = b.identity.cache_key + '.db'
ps.add(message='Upload bundle file', item_type='bytes', item_count=0)
total = [0]
@call_interval(5)
def x(n):
total[0] += n
ps.update(message='Upload bundle file', item_count=total[0])
with open(db_path) as f:
remote.makedir(os.path.dirname(db_ck), recursive=True,
allow_recreate=True)
self.logger.info('Send bundle file {} '.format(db_path))
e = remote.setcontents_async(db_ck, f, progress_callback=upload_cb)
e.wait()
ps.update(state='done')
if not no_partitions:
for p in b.partitions:
ps.add(message='Upload partition', item_type='bytes',
item_count=0, p_vid=p.vid)
with p.datafile.open(mode='rb') as fin:
total = [0]
@call_interval(5)
def x(bytes):
total[0] += bytes
ps.update(message='Upload partition'.format(p.
identity.vname), item_count=total[0])
remote.makedir(os.path.dirname(p.datafile.path),
recursive=True, allow_recreate=True)
event = remote.setcontents_async(p.datafile.path, fin,
progress_callback=progress)
event.wait()
ps.update(state='done')
ps.add(message='Setting metadata')
ident = json.dumps(b.identity.dict)
remote.setcontents(os.path.join('_meta', 'vid', b.identity.vid), ident)
remote.setcontents(os.path.join('_meta', 'id', b.identity.id_), ident)
remote.setcontents(os.path.join('_meta', 'vname', text_type(b.
identity.vname)), ident)
remote.setcontents(os.path.join('_meta', 'name', text_type(b.
identity.name)), ident)
ps.update(state='done')
b.dataset.commit()
return remote_name, db_ck
| send_to_remote | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, b, no_partitions=False):\n raise DeprecationWarning(\"Don't use any more?\")\n from ambry.bundle.process import call_interval\n remote_name = self.resolve_remote(b)\n remote = self.remote(remote_name)\n db_path = b.package()\n with b.progress.start('checkin', 0, message='Check in bundle') as ps:\n ps.add(message='Checking in bundle {} to {}'.format(b.identity.\n vname, remote))\n db_ck = b.identity.cache_key + '.db'\n ps.add(message='Upload bundle file', item_type='bytes', item_count=0)\n total = [0]\n\n @call_interval(5)\n def x(n):\n total[0] += n\n ps.update(message='Upload bundle file', item_count=total[0])\n with open(db_path) as f:\n remote.makedir(os.path.dirname(db_ck), recursive=True,\n allow_recreate=True)\n self.logger.info('Send bundle file {} '.format(db_path))\n e = remote.setcontents_async(db_ck, f, progress_callback=upload_cb)\n e.wait()\n ps.update(state='done')\n if not no_partitions:\n for p in b.partitions:\n ps.add(message='Upload partition', item_type='bytes',\n item_count=0, p_vid=p.vid)\n with p.datafile.open(mode='rb') as fin:\n total = [0]\n\n @call_interval(5)\n def x(bytes):\n total[0] += bytes\n ps.update(message='Upload partition'.format(p.\n identity.vname), item_count=total[0])\n remote.makedir(os.path.dirname(p.datafile.path),\n recursive=True, allow_recreate=True)\n event = remote.setcontents_async(p.datafile.path, fin,\n progress_callback=progress)\n event.wait()\n ps.update(state='done')\n ps.add(message='Setting metadata')\n ident = json.dumps(b.identity.dict)\n remote.setcontents(os.path.join('_meta', 'vid', b.identity.vid), ident)\n remote.setcontents(os.path.join('_meta', 'id', b.identity.id_), ident)\n remote.setcontents(os.path.join('_meta', 'vname', text_type(b.\n identity.vname)), ident)\n remote.setcontents(os.path.join('_meta', 'name', text_type(b.\n identity.name)), ident)\n ps.update(state='done')\n b.dataset.commit()\n return remote_name, db_ck\n",
"role": "user"
},
{
"content": "send_to_remote",
"role": "assistant"
}
] |
def x(self, ref, remote=None):
if not remote:
remote, vname = self.find_remote_bundle(ref)
if vname:
ref = vname
else:
pass
if not remote:
raise NotFoundError("Failed to find bundle ref '{}' in any remote".
format(ref))
self.logger.info("Load '{}' from '{}'".format(ref, remote))
vid = self._checkin_remote_bundle(remote, ref)
self.commit()
return vid
| checkin_remote_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref, remote=None):\n if not remote:\n remote, vname = self.find_remote_bundle(ref)\n if vname:\n ref = vname\n else:\n pass\n if not remote:\n raise NotFoundError(\"Failed to find bundle ref '{}' in any remote\".\n format(ref))\n self.logger.info(\"Load '{}' from '{}'\".format(ref, remote))\n vid = self._checkin_remote_bundle(remote, ref)\n self.commit()\n return vid\n",
"role": "user"
},
{
"content": "checkin_remote_bundle",
"role": "assistant"
}
] |
def x(self, remote, ref):
from ambry.bundle.process import call_interval
from ambry.orm.exc import NotFoundError
from ambry.orm import Remote
from ambry.util.flo import copy_file_or_flo
from tempfile import NamedTemporaryFile
assert isinstance(remote, Remote)
@call_interval(5)
def x(r, total):
self.logger.info('{}: Downloaded {} bytes'.format(ref, total))
b = None
try:
b = self.bundle(ref)
self.logger.info('{}: Already installed'.format(ref))
vid = b.identity.vid
except NotFoundError:
self.logger.info('{}: Syncing'.format(ref))
db_dir = self.filesystem.downloads('bundles')
db_f = os.path.join(db_dir, ref)
if not os.path.exists(os.path.join(db_dir, db_f)):
self.logger.info("Downloading bundle '{}' to '{}".format(ref, db_f)
)
with open(db_f, 'wb') as f_out:
with remote.checkout(ref) as f:
copy_file_or_flo(f, f_out, cb=cb)
f_out.flush()
self.checkin_bundle(db_f)
b = self.bundle(ref)
b.dataset.data['remote_name'] = remote.short_name
b.dataset.upstream = remote.url
b.dstate = b.STATES.CHECKEDOUT
b.commit()
finally:
if b:
b.progress.close()
vid = b.identity.vid
return vid
| checkin_remote_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, remote, ref):\n from ambry.bundle.process import call_interval\n from ambry.orm.exc import NotFoundError\n from ambry.orm import Remote\n from ambry.util.flo import copy_file_or_flo\n from tempfile import NamedTemporaryFile\n assert isinstance(remote, Remote)\n\n @call_interval(5)\n def x(r, total):\n self.logger.info('{}: Downloaded {} bytes'.format(ref, total))\n b = None\n try:\n b = self.bundle(ref)\n self.logger.info('{}: Already installed'.format(ref))\n vid = b.identity.vid\n except NotFoundError:\n self.logger.info('{}: Syncing'.format(ref))\n db_dir = self.filesystem.downloads('bundles')\n db_f = os.path.join(db_dir, ref)\n if not os.path.exists(os.path.join(db_dir, db_f)):\n self.logger.info(\"Downloading bundle '{}' to '{}\".format(ref, db_f)\n )\n with open(db_f, 'wb') as f_out:\n with remote.checkout(ref) as f:\n copy_file_or_flo(f, f_out, cb=cb)\n f_out.flush()\n self.checkin_bundle(db_f)\n b = self.bundle(ref)\n b.dataset.data['remote_name'] = remote.short_name\n b.dataset.upstream = remote.url\n b.dstate = b.STATES.CHECKEDOUT\n b.commit()\n finally:\n if b:\n b.progress.close()\n vid = b.identity.vid\n return vid\n",
"role": "user"
},
{
"content": "checkin_remote_bundle",
"role": "assistant"
}
] |
def x(self):
from ambry.orm import Remote
for r in self.database.session.query(Remote).all():
if not r.short_name:
continue
yield self.remote(r.short_name)
| remotes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.orm import Remote\n for r in self.database.session.query(Remote).all():\n if not r.short_name:\n continue\n yield self.remote(r.short_name)\n",
"role": "user"
},
{
"content": "remotes",
"role": "assistant"
}
] |
def x(self, name):
from ambry.orm import Remote
from sqlalchemy import or_
from ambry.orm.exc import NotFoundError
from sqlalchemy.orm.exc import NoResultFound, MultipleResultsFound
if not name.strip():
raise NotFoundError('Empty remote name')
try:
try:
r = self.database.session.query(Remote).filter(Remote.
short_name == name).one()
except NoResultFound as e:
r = None
if not r:
r = self.database.session.query(Remote).filter(Remote.url == name
).one()
except NoResultFound as e:
raise NotFoundError(str(e) + '; ' + name)
except MultipleResultsFound as e:
self.logger.error("Got multiple results for search for remote '{}': {}"
.format(name, e))
return None
return r
| remote | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, name):\n from ambry.orm import Remote\n from sqlalchemy import or_\n from ambry.orm.exc import NotFoundError\n from sqlalchemy.orm.exc import NoResultFound, MultipleResultsFound\n if not name.strip():\n raise NotFoundError('Empty remote name')\n try:\n try:\n r = self.database.session.query(Remote).filter(Remote.\n short_name == name).one()\n except NoResultFound as e:\n r = None\n if not r:\n r = self.database.session.query(Remote).filter(Remote.url == name\n ).one()\n except NoResultFound as e:\n raise NotFoundError(str(e) + '; ' + name)\n except MultipleResultsFound as e:\n self.logger.error(\"Got multiple results for search for remote '{}': {}\"\n .format(name, e))\n return None\n return r\n",
"role": "user"
},
{
"content": "remote",
"role": "assistant"
}
] |
def x(self, ref, remote_service_type='s3'):
for r in self.remotes:
if remote_service_type and r.service != remote_service_type:
continue
if 'list' not in r.data:
continue
for k, v in r.data['list'].items():
if ref in v.values():
return r, v['vname']
return None, None
| find_remote_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref, remote_service_type='s3'):\n for r in self.remotes:\n if remote_service_type and r.service != remote_service_type:\n continue\n if 'list' not in r.data:\n continue\n for k, v in r.data['list'].items():\n if ref in v.values():\n return r, v['vname']\n return None, None\n",
"role": "user"
},
{
"content": "find_remote_bundle",
"role": "assistant"
}
] |
def x(self, ref, try_harder=None):
from ambry.identity import ObjectNumber
remote, vid = self._find_remote_bundle(ref)
if remote:
return remote, vid
if try_harder:
on = ObjectNumber.parse(vid)
if on:
raise NotImplementedError()
don = on.as_dataset
return self._find_remote_bundle(vid)
parts = ref.split('-')
for i in range(len(parts) - 1, 2, -1):
remote, vid = self._find_remote_bundle('-'.join(parts[:i]))
if remote:
return remote, vid
return None, None
| find_remote_bundle | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ref, try_harder=None):\n from ambry.identity import ObjectNumber\n remote, vid = self._find_remote_bundle(ref)\n if remote:\n return remote, vid\n if try_harder:\n on = ObjectNumber.parse(vid)\n if on:\n raise NotImplementedError()\n don = on.as_dataset\n return self._find_remote_bundle(vid)\n parts = ref.split('-')\n for i in range(len(parts) - 1, 2, -1):\n remote, vid = self._find_remote_bundle('-'.join(parts[:i]))\n if remote:\n return remote, vid\n return None, None\n",
"role": "user"
},
{
"content": "find_remote_bundle",
"role": "assistant"
}
] |
def x(self, url):
from sqlalchemy.orm.exc import NoResultFound
from ambry.orm.exc import NotFoundError
from ambry.util import parse_url_to_dict
from ambry.orm import Account
pd = parse_url_to_dict(url)
try:
act = self.database.session.query(Account).filter(Account.
account_id == pd['netloc']).one()
act.secret_password = self._account_password
return act
except NoResultFound:
pass
for r in self.remotes:
if url.startswith(r.url):
return r
raise NotFoundError("Did not find account for url: '{}' ".format(url))
| account | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, url):\n from sqlalchemy.orm.exc import NoResultFound\n from ambry.orm.exc import NotFoundError\n from ambry.util import parse_url_to_dict\n from ambry.orm import Account\n pd = parse_url_to_dict(url)\n try:\n act = self.database.session.query(Account).filter(Account.\n account_id == pd['netloc']).one()\n act.secret_password = self._account_password\n return act\n except NoResultFound:\n pass\n for r in self.remotes:\n if url.startswith(r.url):\n return r\n raise NotFoundError(\"Did not find account for url: '{}' \".format(url))\n",
"role": "user"
},
{
"content": "account",
"role": "assistant"
}
] |
def x(self):
d = {}
if False and not self._account_password:
from ambry.dbexceptions import ConfigurationError
raise ConfigurationError(
"Can't access accounts without setting an account password either in the accounts.password config, or in the AMBRY_ACCOUNT_PASSWORD env var."
)
for act in self.database.session.query(Account).all():
if self._account_password:
act.secret_password = self._account_password
e = act.dict
a_id = e['account_id']
d[a_id] = e
return d
| accounts | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n d = {}\n if False and not self._account_password:\n from ambry.dbexceptions import ConfigurationError\n raise ConfigurationError(\n \"Can't access accounts without setting an account password either in the accounts.password config, or in the AMBRY_ACCOUNT_PASSWORD env var.\"\n )\n for act in self.database.session.query(Account).all():\n if self._account_password:\n act.secret_password = self._account_password\n e = act.dict\n a_id = e['account_id']\n d[a_id] = e\n return d\n",
"role": "user"
},
{
"content": "accounts",
"role": "assistant"
}
] |
def x(self, assignment_class=None, namespace='d'):
if assignment_class == 'self':
return str(DatasetNumber())
elif assignment_class is None:
try:
nsconfig = self.services['numbers']
except ConfigurationError:
self.logger.error(
'No number server configuration; returning self assigned number'
)
return str(DatasetNumber())
for assignment_class in ('self', 'unregistered', 'registered',
'authority'):
if assignment_class + '-key' in nsconfig:
break
if assignment_class == 'self':
return str(DatasetNumber())
else:
try:
nsconfig = self.services['numbers']
except ConfigurationError:
raise ConfigurationError('No number server configuration')
if assignment_class + '-key' not in nsconfig:
raise ConfigurationError(
'Assignment class {} not number server config'.format(
assignment_class))
try:
key = nsconfig[assignment_class + '-key']
config = {'key': key, 'host': nsconfig['host'], 'port': nsconfig.
get('port', 80)}
ns = NumberServer(**config)
n = str(next(ns))
self.logger.info('Got number from number server: {}'.format(n))
except HTTPError as e:
self.logger.error('Failed to get number from number server for key: {}'
.format(key, e.message))
self.logger.error(
'Using self-generated number. There is no problem with this, but they are longer than centrally generated numbers.'
)
n = str(DatasetNumber())
return n
| number | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, assignment_class=None, namespace='d'):\n if assignment_class == 'self':\n return str(DatasetNumber())\n elif assignment_class is None:\n try:\n nsconfig = self.services['numbers']\n except ConfigurationError:\n self.logger.error(\n 'No number server configuration; returning self assigned number'\n )\n return str(DatasetNumber())\n for assignment_class in ('self', 'unregistered', 'registered',\n 'authority'):\n if assignment_class + '-key' in nsconfig:\n break\n if assignment_class == 'self':\n return str(DatasetNumber())\n else:\n try:\n nsconfig = self.services['numbers']\n except ConfigurationError:\n raise ConfigurationError('No number server configuration')\n if assignment_class + '-key' not in nsconfig:\n raise ConfigurationError(\n 'Assignment class {} not number server config'.format(\n assignment_class))\n try:\n key = nsconfig[assignment_class + '-key']\n config = {'key': key, 'host': nsconfig['host'], 'port': nsconfig.\n get('port', 80)}\n ns = NumberServer(**config)\n n = str(next(ns))\n self.logger.info('Got number from number server: {}'.format(n))\n except HTTPError as e:\n self.logger.error('Failed to get number from number server for key: {}'\n .format(key, e.message))\n self.logger.error(\n 'Using self-generated number. There is no problem with this, but they are longer than centrally generated numbers.'\n )\n n = str(DatasetNumber())\n return n\n",
"role": "user"
},
{
"content": "number",
"role": "assistant"
}
] |
def x(self):
ret = self._db.session.query(Config).filter(Config.type == 'buildstate'
).filter(Config.group == 'access').filter(Config.key == 'last'
).order_by(Config.modified.desc()).all()
return ret
| edit_history | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ret = self._db.session.query(Config).filter(Config.type == 'buildstate'\n ).filter(Config.group == 'access').filter(Config.key == 'last'\n ).order_by(Config.modified.desc()).all()\n return ret\n",
"role": "user"
},
{
"content": "edit_history",
"role": "assistant"
}
] |
def x(self, dir, detach=False, force=False):
import yaml
fs = fsopendir(dir)
bundles = []
for f in fs.walkfiles(wildcard='bundle.yaml'):
self.logger.info('Visiting {}'.format(f))
config = yaml.load(fs.getcontents(f))
if not config:
self.logger.error(
"Failed to get a valid bundle configuration from '{}'".
format(f))
bid = config['identity']['id']
try:
b = self.bundle(bid)
except NotFoundError:
b = None
if not b:
b = self.new_from_bundle_config(config)
self.logger.info('{} Loading New'.format(b.identity.fqname))
else:
self.logger.info('{} Loading Existing'.format(b.identity.fqname))
source_url = os.path.dirname(fs.getsyspath(f))
b.set_file_system(source_url=source_url)
self.logger.info('{} Loading from {}'.format(b.identity.fqname,
source_url))
b.sync_in()
if detach:
self.logger.info('{} Detaching'.format(b.identity.fqname))
b.set_file_system(source_url=None)
if force:
self.logger.info('{} Sync out'.format(b.identity.fqname))
b = self.bundle(bid)
b.sync_out()
bundles.append(b)
b.close()
return bundles
| import_bundles | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, dir, detach=False, force=False):\n import yaml\n fs = fsopendir(dir)\n bundles = []\n for f in fs.walkfiles(wildcard='bundle.yaml'):\n self.logger.info('Visiting {}'.format(f))\n config = yaml.load(fs.getcontents(f))\n if not config:\n self.logger.error(\n \"Failed to get a valid bundle configuration from '{}'\".\n format(f))\n bid = config['identity']['id']\n try:\n b = self.bundle(bid)\n except NotFoundError:\n b = None\n if not b:\n b = self.new_from_bundle_config(config)\n self.logger.info('{} Loading New'.format(b.identity.fqname))\n else:\n self.logger.info('{} Loading Existing'.format(b.identity.fqname))\n source_url = os.path.dirname(fs.getsyspath(f))\n b.set_file_system(source_url=source_url)\n self.logger.info('{} Loading from {}'.format(b.identity.fqname,\n source_url))\n b.sync_in()\n if detach:\n self.logger.info('{} Detaching'.format(b.identity.fqname))\n b.set_file_system(source_url=None)\n if force:\n self.logger.info('{} Sync out'.format(b.identity.fqname))\n b = self.bundle(bid)\n b.sync_out()\n bundles.append(b)\n b.close()\n return bundles\n",
"role": "user"
},
{
"content": "import_bundles",
"role": "assistant"
}
] |
def x(self, limited_run=False):
from multiprocessing import cpu_count
from ambry.bundle.concurrent import Pool, init_library
if self.processes:
cpus = self.processes
else:
cpus = cpu_count()
self.logger.info('Starting MP pool with {} processors'.format(cpus))
return Pool(self, processes=cpus, initializer=init_library,
maxtasksperchild=1, initargs=[self.database.dsn, self.
_account_password, limited_run])
| process_pool | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, limited_run=False):\n from multiprocessing import cpu_count\n from ambry.bundle.concurrent import Pool, init_library\n if self.processes:\n cpus = self.processes\n else:\n cpus = cpu_count()\n self.logger.info('Starting MP pool with {} processors'.format(cpus))\n return Pool(self, processes=cpus, initializer=init_library,\n maxtasksperchild=1, initargs=[self.database.dsn, self.\n _account_password, limited_run])\n",
"role": "user"
},
{
"content": "process_pool",
"role": "assistant"
}
] |
def x():
import sys
import inspect
try:
raise Exception
except:
file_ = '.../' + '/'.join(inspect.currentframe().f_code.co_filename
.split('/')[-3:])
line_ = sys.exc_info()[2].tb_frame.f_back.f_lineno
return '{}:{}'.format(file_, line_)
| file_loc | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x():\n import sys\n import inspect\n try:\n raise Exception\n except:\n file_ = '.../' + '/'.join(inspect.currentframe().f_code.co_filename\n .split('/')[-3:])\n line_ = sys.exc_info()[2].tb_frame.f_back.f_lineno\n return '{}:{}'.format(file_, line_)\n",
"role": "user"
},
{
"content": "file_loc",
"role": "assistant"
}
] |
def x(bundle, source_headers, dest_table, env):
dest_headers = [c.name for c in dest_table.columns]
row_processors = []
out = [file_header]
transforms = list(dest_table.transforms)
column_names = []
column_types = []
for i, segments in enumerate(transforms):
seg_funcs = []
for col_num, (segment, column) in enumerate(zip(segments,
dest_table.columns), 1):
if not segment:
seg_funcs.append('row[{}]'.format(col_num - 1))
continue
assert column
assert column.name == segment['column'].name
col_name = column.name
preamble, try_lines, exception = make_stack(env, i, segment)
assert col_num == column.sequence_id, (dest_table.name, col_num,
column.sequence_id)
column_names.append(col_name)
column_types.append(column.datatype)
f_name = '{table_name}_{column_name}_{stage}'.format(table_name
=dest_table.name, column_name=col_name, stage=i)
exception = (exception if exception else
'raise ValueError("Failed to cast column \'{}\', in function {}, value \'{}\': {}".format(header_d,"'
+ f_name + '", v.encode(\'ascii\', \'replace\'), exc) ) ')
try:
i_s = source_headers.index(column.name)
header_s = column.name
v = 'row[{}]'.format(i_s)
except ValueError as e:
i_s = 'None'
header_s = None
v = 'None' if col_num > 1 else 'row_n'
i_d = column.sequence_id - 1
header_d = column.name
template_args = dict(f_name=f_name, table_name=dest_table.name,
column_name=col_name, stage=i, i_s=i_s, i_d=i_d, header_s=
header_s, header_d=header_d, v=v, exception=indent +
exception, stack='\n'.join(indent + l for l in try_lines),
col_args='# col_args not implemented yet')
seg_funcs.append(f_name +
"({v}, {i_s}, {i_d}, {header_s}, '{header_d}', row, row_n, errors, scratch, accumulator, pipe, bundle, source)"
.format(v=v, i_s=i_s, i_d=i_d, header_s="'" + header_s +
"'" if header_s else 'None', header_d=header_d))
out.append('\n'.join(preamble))
out.append(column_template.format(**template_args))
source_headers = dest_headers
stack = '\n'.join('{}{}, # {}'.format(indent, l, cn) for l, cn, dt in
zip(seg_funcs, column_names, column_types))
out.append(row_template.format(table=dest_table.name, stage=i,
stack=stack))
row_processors.append('row_{table}_{stage}'.format(stage=i, table=
dest_table.name))
stack = '\n'.join("{}cast_{}(row[{}], '{}', errors),".format(indent, c.
datatype, i, c.name) for i, c in enumerate(dest_table.columns))
out.append(row_template.format(table=dest_table.name, stage=len(
transforms), stack=stack))
row_processors.append('row_{table}_{stage}'.format(stage=len(transforms
), table=dest_table.name))
out.append('row_processors = [{}]'.format(','.join(row_processors)))
return '\n'.join(out)
| make_row_processors | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(bundle, source_headers, dest_table, env):\n dest_headers = [c.name for c in dest_table.columns]\n row_processors = []\n out = [file_header]\n transforms = list(dest_table.transforms)\n column_names = []\n column_types = []\n for i, segments in enumerate(transforms):\n seg_funcs = []\n for col_num, (segment, column) in enumerate(zip(segments,\n dest_table.columns), 1):\n if not segment:\n seg_funcs.append('row[{}]'.format(col_num - 1))\n continue\n assert column\n assert column.name == segment['column'].name\n col_name = column.name\n preamble, try_lines, exception = make_stack(env, i, segment)\n assert col_num == column.sequence_id, (dest_table.name, col_num,\n column.sequence_id)\n column_names.append(col_name)\n column_types.append(column.datatype)\n f_name = '{table_name}_{column_name}_{stage}'.format(table_name\n =dest_table.name, column_name=col_name, stage=i)\n exception = (exception if exception else \n 'raise ValueError(\"Failed to cast column \\'{}\\', in function {}, value \\'{}\\': {}\".format(header_d,\"'\n + f_name + '\", v.encode(\\'ascii\\', \\'replace\\'), exc) ) ')\n try:\n i_s = source_headers.index(column.name)\n header_s = column.name\n v = 'row[{}]'.format(i_s)\n except ValueError as e:\n i_s = 'None'\n header_s = None\n v = 'None' if col_num > 1 else 'row_n'\n i_d = column.sequence_id - 1\n header_d = column.name\n template_args = dict(f_name=f_name, table_name=dest_table.name,\n column_name=col_name, stage=i, i_s=i_s, i_d=i_d, header_s=\n header_s, header_d=header_d, v=v, exception=indent +\n exception, stack='\\n'.join(indent + l for l in try_lines),\n col_args='# col_args not implemented yet')\n seg_funcs.append(f_name +\n \"({v}, {i_s}, {i_d}, {header_s}, '{header_d}', row, row_n, errors, scratch, accumulator, pipe, bundle, source)\"\n .format(v=v, i_s=i_s, i_d=i_d, header_s=\"'\" + header_s +\n \"'\" if header_s else 'None', header_d=header_d))\n out.append('\\n'.join(preamble))\n out.append(column_template.format(**template_args))\n source_headers = dest_headers\n stack = '\\n'.join('{}{}, # {}'.format(indent, l, cn) for l, cn, dt in\n zip(seg_funcs, column_names, column_types))\n out.append(row_template.format(table=dest_table.name, stage=i,\n stack=stack))\n row_processors.append('row_{table}_{stage}'.format(stage=i, table=\n dest_table.name))\n stack = '\\n'.join(\"{}cast_{}(row[{}], '{}', errors),\".format(indent, c.\n datatype, i, c.name) for i, c in enumerate(dest_table.columns))\n out.append(row_template.format(table=dest_table.name, stage=len(\n transforms), stack=stack))\n row_processors.append('row_{table}_{stage}'.format(stage=len(transforms\n ), table=dest_table.name))\n out.append('row_processors = [{}]'.format(','.join(row_processors)))\n return '\\n'.join(out)\n",
"role": "user"
},
{
"content": "make_row_processors",
"role": "assistant"
}
] |
def x(f, f_name=None, raise_for_missing=True):
import inspect
from ambry.dbexceptions import ConfigurationError
if inspect.isclass(f):
try:
args = inspect.getargspec(f.__init__).args
except TypeError as e:
raise TypeError('Failed to inspect {}: {}'.format(f, e))
else:
args = inspect.getargspec(f).args
if len(args) > 1 and args[0] == 'self':
args = args[1:]
for a in args:
if a not in all_args + ('exception',):
if raise_for_missing:
raise ConfigurationError(
"Caster code {} has unknown argument name: '{}'. Must be one of: {} "
.format(f, a, ','.join(all_args)))
arg_map = {e: e for e in var_args}
args = [arg_map.get(a, a) for a in args]
return '{}({})'.format(f_name if f_name else f.__name__, ','.join(args))
| calling_code | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(f, f_name=None, raise_for_missing=True):\n import inspect\n from ambry.dbexceptions import ConfigurationError\n if inspect.isclass(f):\n try:\n args = inspect.getargspec(f.__init__).args\n except TypeError as e:\n raise TypeError('Failed to inspect {}: {}'.format(f, e))\n else:\n args = inspect.getargspec(f).args\n if len(args) > 1 and args[0] == 'self':\n args = args[1:]\n for a in args:\n if a not in all_args + ('exception',):\n if raise_for_missing:\n raise ConfigurationError(\n \"Caster code {} has unknown argument name: '{}'. Must be one of: {} \"\n .format(f, a, ','.join(all_args)))\n arg_map = {e: e for e in var_args}\n args = [arg_map.get(a, a) for a in args]\n return '{}({})'.format(f_name if f_name else f.__name__, ','.join(args))\n",
"role": "user"
},
{
"content": "calling_code",
"role": "assistant"
}
] |
def x(env, stage, segment):
import string
import random
from ambry.valuetype import ValueType
column = segment['column']
def x(column, t):
preamble = []
line_t = 'v = {} # {}'
if isinstance(t, type) and issubclass(t, ValueType):
try:
cc, fl = calling_code(t, t.__name__), file_loc()
except TypeError:
cc, fl = '{}(v)'.format(t.__name__), file_loc()
preamble.append("{} = resolve_value_type('{}') # {}".format(t.
__name__, t.vt_code, fl))
elif isinstance(t, type):
cc, fl = 'parse_{}(v, header_d)'.format(t.__name__), file_loc()
elif callable(env.get(t)):
cc, fl = calling_code(env.get(t), t), file_loc()
else:
rnd = ''.join(random.choice(string.ascii_lowercase) for _ in
range(6))
name = 'tg_{}_{}_{}'.format(column.name, stage, rnd)
try:
a, b, fl = rewrite_tg(env, name, t)
except CodeGenError as e:
raise CodeGenError(
"Failed to re-write pipe code '{}' in column '{}.{}': {} "
.format(t, column.table.name, column.name, e))
cc = str(a)
if b:
preamble.append('{} = {} # {}'.format(name, b, fl))
line = line_t.format(cc, fl)
return line, preamble
preamble = []
try_lines = []
for t in ([segment['init'], segment['datatype']] + segment['transforms']):
if not t:
continue
line, col_preamble = make_line(column, t)
preamble += col_preamble
try_lines.append(line)
exception = None
if segment['exception']:
exception, col_preamble = make_line(column, segment['exception'])
if len(try_lines) == 0:
try_lines.append('pass # Empty pipe segment')
assert len(try_lines) > 0, column.name
return preamble, try_lines, exception
| make_stack | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(env, stage, segment):\n import string\n import random\n from ambry.valuetype import ValueType\n column = segment['column']\n\n def x(column, t):\n preamble = []\n line_t = 'v = {} # {}'\n if isinstance(t, type) and issubclass(t, ValueType):\n try:\n cc, fl = calling_code(t, t.__name__), file_loc()\n except TypeError:\n cc, fl = '{}(v)'.format(t.__name__), file_loc()\n preamble.append(\"{} = resolve_value_type('{}') # {}\".format(t.\n __name__, t.vt_code, fl))\n elif isinstance(t, type):\n cc, fl = 'parse_{}(v, header_d)'.format(t.__name__), file_loc()\n elif callable(env.get(t)):\n cc, fl = calling_code(env.get(t), t), file_loc()\n else:\n rnd = ''.join(random.choice(string.ascii_lowercase) for _ in\n range(6))\n name = 'tg_{}_{}_{}'.format(column.name, stage, rnd)\n try:\n a, b, fl = rewrite_tg(env, name, t)\n except CodeGenError as e:\n raise CodeGenError(\n \"Failed to re-write pipe code '{}' in column '{}.{}': {} \"\n .format(t, column.table.name, column.name, e))\n cc = str(a)\n if b:\n preamble.append('{} = {} # {}'.format(name, b, fl))\n line = line_t.format(cc, fl)\n return line, preamble\n preamble = []\n try_lines = []\n for t in ([segment['init'], segment['datatype']] + segment['transforms']):\n if not t:\n continue\n line, col_preamble = make_line(column, t)\n preamble += col_preamble\n try_lines.append(line)\n exception = None\n if segment['exception']:\n exception, col_preamble = make_line(column, segment['exception'])\n if len(try_lines) == 0:\n try_lines.append('pass # Empty pipe segment')\n assert len(try_lines) > 0, column.name\n return preamble, try_lines, exception\n",
"role": "user"
},
{
"content": "make_stack",
"role": "assistant"
}
] |
def x(env, tg_name, code):
visitor = ReplaceTG(env, tg_name)
assert visitor.tg_name
tree = visitor.visit(ast.parse(code))
if visitor.loc:
loc = ' #' + visitor.loc
else:
loc = file_loc()
if visitor.trans_gen:
tg = meta.dump_python_source(visitor.trans_gen).strip()
else:
tg = None
return meta.dump_python_source(tree).strip(), tg, loc
| rewrite_tg | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(env, tg_name, code):\n visitor = ReplaceTG(env, tg_name)\n assert visitor.tg_name\n tree = visitor.visit(ast.parse(code))\n if visitor.loc:\n loc = ' #' + visitor.loc\n else:\n loc = file_loc()\n if visitor.trans_gen:\n tg = meta.dump_python_source(visitor.trans_gen).strip()\n else:\n tg = None\n return meta.dump_python_source(tree).strip(), tg, loc\n",
"role": "user"
},
{
"content": "rewrite_tg",
"role": "assistant"
}
] |
def x(self, el):
count = next(self.counter)
heapq.heappush(self._queue, (el, count))
| push | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, el):\n count = next(self.counter)\n heapq.heappush(self._queue, (el, count))\n",
"role": "user"
},
{
"content": "push",
"role": "assistant"
}
] |
def x(self):
sc = self._p.space_coverage
gc = self._p.grain_coverage
if sc and gc:
if parse_to_gvid(gc[0]).level == 'state' and parse_to_gvid(sc[0]
).level == 'state':
return parse_to_gvid(sc[0]).geo_name
else:
return '{} in {}'.format(parse_to_gvid(gc[0]).level_plural.
title(), parse_to_gvid(sc[0]).geo_name)
elif sc:
return parse_to_gvid(sc[0]).geo_name.title()
elif sc:
return parse_to_gvid(gc[0]).level_plural.title()
else:
return ''
| geo_description | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n sc = self._p.space_coverage\n gc = self._p.grain_coverage\n if sc and gc:\n if parse_to_gvid(gc[0]).level == 'state' and parse_to_gvid(sc[0]\n ).level == 'state':\n return parse_to_gvid(sc[0]).geo_name\n else:\n return '{} in {}'.format(parse_to_gvid(gc[0]).level_plural.\n title(), parse_to_gvid(sc[0]).geo_name)\n elif sc:\n return parse_to_gvid(sc[0]).geo_name.title()\n elif sc:\n return parse_to_gvid(gc[0]).level_plural.title()\n else:\n return ''\n",
"role": "user"
},
{
"content": "geo_description",
"role": "assistant"
}
] |
def x(self):
tc = [t for t in self._p.time_coverage if t]
if not tc:
return ''
mn = min(tc)
mx = max(tc)
if not mn and not mx:
return ''
elif mn == mx:
return mn
else:
return '{} to {}'.format(mn, mx)
| time_description | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n tc = [t for t in self._p.time_coverage if t]\n if not tc:\n return ''\n mn = min(tc)\n mx = max(tc)\n if not mn and not mx:\n return ''\n elif mn == mx:\n return mn\n else:\n return '{} to {}'.format(mn, mx)\n",
"role": "user"
},
{
"content": "time_description",
"role": "assistant"
}
] |
def x(self):
gd = self.geo_description
td = self.time_description
if gd and td:
return '{}, {}. {} Rows.'.format(gd, td, self._p.count)
elif gd:
return '{}. {} Rows.'.format(gd, self._p.count)
elif td:
return '{}. {} Rows.'.format(td, self._p.count)
else:
return '{} Rows.'.format(self._p.count)
| sub_description | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n gd = self.geo_description\n td = self.time_description\n if gd and td:\n return '{}, {}. {} Rows.'.format(gd, td, self._p.count)\n elif gd:\n return '{}. {} Rows.'.format(gd, self._p.count)\n elif td:\n return '{}. {} Rows.'.format(td, self._p.count)\n else:\n return '{} Rows.'.format(self._p.count)\n",
"role": "user"
},
{
"content": "sub_description",
"role": "assistant"
}
] |
def x(self):
if self.dataset is None:
s = object_session(self)
ds = s.query(Dataset).filter(Dataset.id_ == self.d_id).one()
else:
ds = self.dataset
d = {'id': self.id, 'vid': self.vid, 'name': self.name, 'vname': self.
vname, 'ref': self.ref, 'space': self.space, 'time': self.time,
'table': self.table_name, 'grain': self.grain, 'variant': self.
variant, 'segment': self.segment, 'format': self.format if self.
format else 'db'}
return PartitionIdentity.from_dict(dict(list(ds.dict.items()) + list(d.
items())))
| identity | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.dataset is None:\n s = object_session(self)\n ds = s.query(Dataset).filter(Dataset.id_ == self.d_id).one()\n else:\n ds = self.dataset\n d = {'id': self.id, 'vid': self.vid, 'name': self.name, 'vname': self.\n vname, 'ref': self.ref, 'space': self.space, 'time': self.time,\n 'table': self.table_name, 'grain': self.grain, 'variant': self.\n variant, 'segment': self.segment, 'format': self.format if self.\n format else 'db'}\n return PartitionIdentity.from_dict(dict(list(ds.dict.items()) + list(d.\n items())))\n",
"role": "user"
},
{
"content": "identity",
"role": "assistant"
}
] |
def x(self, stats):
from ambry.util.datestimes import expand_to_years
scov = set()
tcov = set()
grains = set()
def x(gvid):
try:
return parse_to_gvid(gvid).summarize()
except:
return None
def x(values, column):
parsed = []
for gvid in values:
if gvid is None or gvid == 'None':
continue
try:
parsed.append(parse_to_gvid(gvid))
except ValueError as e:
if self._bundle:
self._bundle.warn(
'While analyzing geo coverage in final partition stage, '
+ "Failed to parse gvid '{}' in {}.{}: {}".format(
str(gvid), column.table.name, column.name, e))
try:
return isimplify(parsed)
except:
return None
def x(year):
try:
return int(year)
except:
return None
for c in self.table.columns:
if c.name not in stats:
continue
try:
if stats[c.name].is_gvid or stats[c.name].is_geoid:
scov |= set(x for x in simplifiy_maybe(stats[c.name].
uniques, c))
grains |= set(summarize_maybe(gvid) for gvid in stats[c.
name].uniques)
elif stats[c.name].is_year:
tcov |= set(int_maybe(x) for x in stats[c.name].uniques)
elif stats[c.name].is_date:
try:
tcov |= set(parser.parse(x, fuzzy=True).year if
isinstance(x, string_types) else x.year for x in
stats[c.name].uniques)
except ValueError:
pass
except Exception as e:
self._bundle.error(
"Failed to set coverage for column '{}', partition '{}': {}"
.format(c.name, self.identity.vname, e))
raise
if 'source_data' in self.data:
for source_name, source in list(self.data['source_data'].items()):
scov.add(self.parse_gvid_or_place(source['space']))
if self.identity.space:
try:
scov.add(self.parse_gvid_or_place(self.identity.space))
except ValueError:
pass
self.space_coverage = sorted([str(x) for x in scov if bool(x) and x.sl in
(10, 40, 50, 60, 160, 400)])
if 'source_data' in self.data:
for source_name, source in list(self.data['source_data'].items()):
if 'time' in source:
for year in expand_to_years(source['time']):
if year:
tcov.add(year)
if self.identity.name.time:
for year in expand_to_years(self.identity.name.time):
if year:
tcov.add(year)
self.time_coverage = [t for t in tcov if t]
if 'source_data' in self.data:
for source_name, source in list(self.data['source_data'].items()):
if 'grain' in source:
grains.add(source['grain'])
self.grain_coverage = sorted(str(g) for g in grains if g)
| set_coverage | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, stats):\n from ambry.util.datestimes import expand_to_years\n scov = set()\n tcov = set()\n grains = set()\n\n def x(gvid):\n try:\n return parse_to_gvid(gvid).summarize()\n except:\n return None\n\n def x(values, column):\n parsed = []\n for gvid in values:\n if gvid is None or gvid == 'None':\n continue\n try:\n parsed.append(parse_to_gvid(gvid))\n except ValueError as e:\n if self._bundle:\n self._bundle.warn(\n 'While analyzing geo coverage in final partition stage, '\n + \"Failed to parse gvid '{}' in {}.{}: {}\".format(\n str(gvid), column.table.name, column.name, e))\n try:\n return isimplify(parsed)\n except:\n return None\n\n def x(year):\n try:\n return int(year)\n except:\n return None\n for c in self.table.columns:\n if c.name not in stats:\n continue\n try:\n if stats[c.name].is_gvid or stats[c.name].is_geoid:\n scov |= set(x for x in simplifiy_maybe(stats[c.name].\n uniques, c))\n grains |= set(summarize_maybe(gvid) for gvid in stats[c.\n name].uniques)\n elif stats[c.name].is_year:\n tcov |= set(int_maybe(x) for x in stats[c.name].uniques)\n elif stats[c.name].is_date:\n try:\n tcov |= set(parser.parse(x, fuzzy=True).year if\n isinstance(x, string_types) else x.year for x in\n stats[c.name].uniques)\n except ValueError:\n pass\n except Exception as e:\n self._bundle.error(\n \"Failed to set coverage for column '{}', partition '{}': {}\"\n .format(c.name, self.identity.vname, e))\n raise\n if 'source_data' in self.data:\n for source_name, source in list(self.data['source_data'].items()):\n scov.add(self.parse_gvid_or_place(source['space']))\n if self.identity.space:\n try:\n scov.add(self.parse_gvid_or_place(self.identity.space))\n except ValueError:\n pass\n self.space_coverage = sorted([str(x) for x in scov if bool(x) and x.sl in\n (10, 40, 50, 60, 160, 400)])\n if 'source_data' in self.data:\n for source_name, source in list(self.data['source_data'].items()):\n if 'time' in source:\n for year in expand_to_years(source['time']):\n if year:\n tcov.add(year)\n if self.identity.name.time:\n for year in expand_to_years(self.identity.name.time):\n if year:\n tcov.add(year)\n self.time_coverage = [t for t in tcov if t]\n if 'source_data' in self.data:\n for source_name, source in list(self.data['source_data'].items()):\n if 'grain' in source:\n grains.add(source['grain'])\n self.grain_coverage = sorted(str(g) for g in grains if g)\n",
"role": "user"
},
{
"content": "set_coverage",
"role": "assistant"
}
] |
def x(self):
d = self.dict
def x(c):
d = c.dict
d['stats'] = [s.dict for s in c.stats]
return d
d['table'] = self.table.dict
d['table']['columns'] = [aug_col(c) for c in self.table.columns]
return d
| detail_dict | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n d = self.dict\n\n def x(c):\n d = c.dict\n d['stats'] = [s.dict for s in c.stats]\n return d\n d['table'] = self.table.dict\n d['table']['columns'] = [aug_col(c) for c in self.table.columns]\n return d\n",
"role": "user"
},
{
"content": "detail_dict",
"role": "assistant"
}
] |
def x(self):
from ambry_sources import MPRowsFile
from fs.errors import ResourceNotFoundError
from ambry.orm.exc import NotFoundError
try:
return MPRowsFile(self._bundle.build_fs, self.cache_key)
except ResourceNotFoundError:
raise NotFoundError(
'Could not locate data file for partition {} (local)'.format(
self.identity.fqname))
| local_datafile | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry_sources import MPRowsFile\n from fs.errors import ResourceNotFoundError\n from ambry.orm.exc import NotFoundError\n try:\n return MPRowsFile(self._bundle.build_fs, self.cache_key)\n except ResourceNotFoundError:\n raise NotFoundError(\n 'Could not locate data file for partition {} (local)'.format(\n self.identity.fqname))\n",
"role": "user"
},
{
"content": "local_datafile",
"role": "assistant"
}
] |
def x(self):
from ambry.exc import NotFoundError
ds = self.dataset
if 'remote_name' not in ds.data:
raise NotFoundError('Could not determine remote for partition: {}'.
format(self.identity.fqname))
return self._bundle.library.remote(ds.data['remote_name'])
| remote | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.exc import NotFoundError\n ds = self.dataset\n if 'remote_name' not in ds.data:\n raise NotFoundError('Could not determine remote for partition: {}'.\n format(self.identity.fqname))\n return self._bundle.library.remote(ds.data['remote_name'])\n",
"role": "user"
},
{
"content": "remote",
"role": "assistant"
}
] |
def x(self):
from ambry.orm.exc import NotFoundError
try:
if self.local_datafile.exists:
return True
except NotFoundError:
pass
return False
| is_local | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.orm.exc import NotFoundError\n try:\n if self.local_datafile.exists:\n return True\n except NotFoundError:\n pass\n return False\n",
"role": "user"
},
{
"content": "is_local",
"role": "assistant"
}
] |
def x(self, ps=None):
from filelock import FileLock
from ambry.util import ensure_dir_exists
from ambry_sources import MPRowsFile
from fs.errors import ResourceNotFoundError
if self.is_local:
return
local = self._bundle.build_fs
b = self._bundle.library.bundle(self.identity.as_dataset().vid)
remote = self._bundle.library.remote(b)
lock_path = local.getsyspath(self.cache_key + '.lock')
ensure_dir_exists(lock_path)
lock = FileLock(lock_path)
if ps:
ps.add_update(message='Localizing {}'.format(self.identity.name),
partition=self, item_type='bytes', state='downloading')
if ps:
def x(bts):
if ps.rec.item_total is None:
ps.rec.item_count = 0
if not ps.rec.data:
ps.rec.data = {}
return self
item_count = ps.rec.item_count + bts
ps.rec.data['updates'] = ps.rec.data.get('updates', 0) + 1
if ps.rec.data['updates'] % 32 == 1:
ps.update(message='Localizing {}'.format(self.identity.name
), item_count=item_count)
else:
from ambry.bundle.process import call_interval
@call_interval(5)
def x(bts):
self._bundle.log('Localizing {}. {} bytes downloaded'.format(
self.vname, bts))
def x(e):
raise e
with lock:
if self.is_local:
return self
try:
with remote.fs.open(self.cache_key + MPRowsFile.EXTENSION, 'rb'
) as f:
event = local.setcontents_async(self.cache_key + MPRowsFile
.EXTENSION, f, progress_callback=progress,
error_callback=exception_cb)
event.wait()
if ps:
ps.update_done()
except ResourceNotFoundError as e:
from ambry.orm.exc import NotFoundError
raise NotFoundError("Failed to get MPRfile '{}' from {}: {} ".
format(self.cache_key, remote.fs, e))
return self
| localize | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, ps=None):\n from filelock import FileLock\n from ambry.util import ensure_dir_exists\n from ambry_sources import MPRowsFile\n from fs.errors import ResourceNotFoundError\n if self.is_local:\n return\n local = self._bundle.build_fs\n b = self._bundle.library.bundle(self.identity.as_dataset().vid)\n remote = self._bundle.library.remote(b)\n lock_path = local.getsyspath(self.cache_key + '.lock')\n ensure_dir_exists(lock_path)\n lock = FileLock(lock_path)\n if ps:\n ps.add_update(message='Localizing {}'.format(self.identity.name),\n partition=self, item_type='bytes', state='downloading')\n if ps:\n\n def x(bts):\n if ps.rec.item_total is None:\n ps.rec.item_count = 0\n if not ps.rec.data:\n ps.rec.data = {}\n return self\n item_count = ps.rec.item_count + bts\n ps.rec.data['updates'] = ps.rec.data.get('updates', 0) + 1\n if ps.rec.data['updates'] % 32 == 1:\n ps.update(message='Localizing {}'.format(self.identity.name\n ), item_count=item_count)\n else:\n from ambry.bundle.process import call_interval\n\n @call_interval(5)\n def x(bts):\n self._bundle.log('Localizing {}. {} bytes downloaded'.format(\n self.vname, bts))\n\n def x(e):\n raise e\n with lock:\n if self.is_local:\n return self\n try:\n with remote.fs.open(self.cache_key + MPRowsFile.EXTENSION, 'rb'\n ) as f:\n event = local.setcontents_async(self.cache_key + MPRowsFile\n .EXTENSION, f, progress_callback=progress,\n error_callback=exception_cb)\n event.wait()\n if ps:\n ps.update_done()\n except ResourceNotFoundError as e:\n from ambry.orm.exc import NotFoundError\n raise NotFoundError(\"Failed to get MPRfile '{}' from {}: {} \".\n format(self.cache_key, remote.fs, e))\n return self\n",
"role": "user"
},
{
"content": "localize",
"role": "assistant"
}
] |
def x(self):
from ambry.orm.exc import NotFoundError
from fs.errors import ResourceNotFoundError
"""The reader for the datafile"""
try:
return self.datafile.reader
except ResourceNotFoundError:
raise NotFoundError("Failed to find partition file, '{}' ".format(
self.datafile.path))
| reader | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.orm.exc import NotFoundError\n from fs.errors import ResourceNotFoundError\n \"\"\"The reader for the datafile\"\"\"\n try:\n return self.datafile.reader\n except ResourceNotFoundError:\n raise NotFoundError(\"Failed to find partition file, '{}' \".format(\n self.datafile.path))\n",
"role": "user"
},
{
"content": "reader",
"role": "assistant"
}
] |
def x(self, predicate=None, headers=None):
with self.reader as r:
for row in r.select(predicate, headers):
yield row
| select | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, predicate=None, headers=None):\n with self.reader as r:\n for row in r.select(predicate, headers):\n yield row\n",
"role": "user"
},
{
"content": "select",
"role": "assistant"
}
] |
def x(self):
if isinstance(self, PartitionProxy):
return AnalysisPartition(self._obj)
else:
return AnalysisPartition(self)
| analysis | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if isinstance(self, PartitionProxy):\n return AnalysisPartition(self._obj)\n else:\n return AnalysisPartition(self)\n",
"role": "user"
},
{
"content": "analysis",
"role": "assistant"
}
] |
def x(self):
if isinstance(self, PartitionProxy):
return MeasureDimensionPartition(self._obj)
else:
return MeasureDimensionPartition(self)
| measuredim | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if isinstance(self, PartitionProxy):\n return MeasureDimensionPartition(self._obj)\n else:\n return MeasureDimensionPartition(self)\n",
"role": "user"
},
{
"content": "measuredim",
"role": "assistant"
}
] |
def x(self, sequence_id=None):
if sequence_id:
self.sequence_id = sequence_id
self._set_ids(force=True)
if self.dataset:
self._update_names()
| update_id | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sequence_id=None):\n if sequence_id:\n self.sequence_id = sequence_id\n self._set_ids(force=True)\n if self.dataset:\n self._update_names()\n",
"role": "user"
},
{
"content": "update_id",
"role": "assistant"
}
] |
def x(self):
d = dict(table=self.table_name, time=self.time, space=self.space, grain
=self.grain, variant=self.variant, segment=self.segment)
assert self.dataset
name = PartialPartitionName(**d).promote(self.dataset.identity.name)
self.name = str(name.name)
self.vname = str(name.vname)
self.cache_key = name.cache_key
self.fqname = str(self.identity.fqname)
| update_names | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n d = dict(table=self.table_name, time=self.time, space=self.space, grain\n =self.grain, variant=self.variant, segment=self.segment)\n assert self.dataset\n name = PartialPartitionName(**d).promote(self.dataset.identity.name)\n self.name = str(name.name)\n self.vname = str(name.vname)\n self.cache_key = name.cache_key\n self.fqname = str(self.identity.fqname)\n",
"role": "user"
},
{
"content": "update_names",
"role": "assistant"
}
] |
def x(mapper, conn, target):
target._set_ids()
if (target.name and target.vname and target.cache_key and target.fqname and
not target.dataset):
return
Partition.before_update(mapper, conn, target)
| before_insert | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(mapper, conn, target):\n target._set_ids()\n if (target.name and target.vname and target.cache_key and target.fqname and\n not target.dataset):\n return\n Partition.before_update(mapper, conn, target)\n",
"role": "user"
},
{
"content": "before_insert",
"role": "assistant"
}
] |
def x(self, predicate=None, filtered_columns=None, columns=None, df_class=None
):
from operator import itemgetter
from ambry.pands import AmbryDataFrame
df_class = df_class or AmbryDataFrame
if columns:
ig = itemgetter(*columns)
else:
ig = None
columns = self.table.header
if filtered_columns:
def x(v):
from six import string_types
if isinstance(v, string_types):
return '"{}"'.format(v)
else:
return v
code = ' and '.join('row.{} == {}'.format(k, maybe_quote(v)) for k,
v in filtered_columns.items())
predicate = eval('lambda row: {}'.format(code))
if predicate:
def x():
for row in self.reader:
if predicate(row):
if ig:
yield ig(row)
else:
yield row.dict
df = df_class(yielder(), columns=columns, partition=self.measuredim)
return df
else:
def x():
for row in self.reader:
yield row.values()
columns = [c for c in self.table.header if c in columns]
return df_class(yielder(), columns=columns, partition=self.measuredim)
| dataframe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, predicate=None, filtered_columns=None, columns=None, df_class=None\n ):\n from operator import itemgetter\n from ambry.pands import AmbryDataFrame\n df_class = df_class or AmbryDataFrame\n if columns:\n ig = itemgetter(*columns)\n else:\n ig = None\n columns = self.table.header\n if filtered_columns:\n\n def x(v):\n from six import string_types\n if isinstance(v, string_types):\n return '\"{}\"'.format(v)\n else:\n return v\n code = ' and '.join('row.{} == {}'.format(k, maybe_quote(v)) for k,\n v in filtered_columns.items())\n predicate = eval('lambda row: {}'.format(code))\n if predicate:\n\n def x():\n for row in self.reader:\n if predicate(row):\n if ig:\n yield ig(row)\n else:\n yield row.dict\n df = df_class(yielder(), columns=columns, partition=self.measuredim)\n return df\n else:\n\n def x():\n for row in self.reader:\n yield row.values()\n columns = [c for c in self.table.header if c in columns]\n return df_class(yielder(), columns=columns, partition=self.measuredim)\n",
"role": "user"
},
{
"content": "dataframe",
"role": "assistant"
}
] |
def x(self, simplify=None, predicate=None):
from shapely.wkt import loads
if not predicate:
predicate = lambda row: True
if simplify:
return [loads(row.geometry).simplify(simplify) for row in self if
predicate(row)]
else:
return [loads(row.geometry) for row in self if predicate(row)]
| shapes | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, simplify=None, predicate=None):\n from shapely.wkt import loads\n if not predicate:\n predicate = lambda row: True\n if simplify:\n return [loads(row.geometry).simplify(simplify) for row in self if\n predicate(row)]\n else:\n return [loads(row.geometry) for row in self if predicate(row)]\n",
"role": "user"
},
{
"content": "shapes",
"role": "assistant"
}
] |
def x(self, basemap, simplify=None, predicate=None, args_f=None, **kwargs):
from descartes import PolygonPatch
from shapely.wkt import loads
from shapely.ops import transform
if not predicate:
predicate = lambda row: True
def x(x, y, z=None):
return basemap(x, y)
def x(shape, row):
args = dict(kwargs.items())
if args_f:
args.update(args_f(row))
return PolygonPatch(transform(map_xform, shape), **args)
def x(row):
if simplify:
shape = loads(row.geometry).simplify(simplify)
else:
shape = loads(row.geometry)
if shape.geom_type == 'MultiPolygon':
for subshape in shape.geoms:
yield make_patch(subshape, row)
else:
yield make_patch(shape, row)
return [patch for row in self if predicate(row) for patch in
yield_patches(row)]
| patches | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, basemap, simplify=None, predicate=None, args_f=None, **kwargs):\n from descartes import PolygonPatch\n from shapely.wkt import loads\n from shapely.ops import transform\n if not predicate:\n predicate = lambda row: True\n\n def x(x, y, z=None):\n return basemap(x, y)\n\n def x(shape, row):\n args = dict(kwargs.items())\n if args_f:\n args.update(args_f(row))\n return PolygonPatch(transform(map_xform, shape), **args)\n\n def x(row):\n if simplify:\n shape = loads(row.geometry).simplify(simplify)\n else:\n shape = loads(row.geometry)\n if shape.geom_type == 'MultiPolygon':\n for subshape in shape.geoms:\n yield make_patch(subshape, row)\n else:\n yield make_patch(shape, row)\n return [patch for row in self if predicate(row) for patch in\n yield_patches(row)]\n",
"role": "user"
},
{
"content": "patches",
"role": "assistant"
}
] |
def x(self):
from ambry.valuetype.core import ROLE
return [c for c in self.columns if c.role == ROLE.MEASURE]
| measures | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from ambry.valuetype.core import ROLE\n return [c for c in self.columns if c.role == ROLE.MEASURE]\n",
"role": "user"
},
{
"content": "measures",
"role": "assistant"
}
] |
def x(self, vid):
from ambry.orm import Column
if isinstance(vid, PartitionColumn):
return vid
elif isinstance(vid, Column):
return PartitionColumn(vid)
else:
return PartitionColumn(self.table.column(vid), self)
| measure | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, vid):\n from ambry.orm import Column\n if isinstance(vid, PartitionColumn):\n return vid\n elif isinstance(vid, Column):\n return PartitionColumn(vid)\n else:\n return PartitionColumn(self.table.column(vid), self)\n",
"role": "user"
},
{
"content": "measure",
"role": "assistant"
}
] |
def x(self, measure, p_dim, s_dim=None, filters={}, df_class=None):
import numpy as np
measure = self.measure(measure)
p_dim = self.dimension(p_dim)
assert p_dim
if s_dim:
s_dim = self.dimension(s_dim)
columns = set([measure.name, p_dim.name])
if p_dim.label:
if p_dim.geoid:
columns.add(p_dim.geoid.name)
columns.add(p_dim.label.name)
if s_dim:
columns.add(s_dim.name)
if s_dim.label:
columns.add(s_dim.label.name)
def x(v):
from six import string_types
if isinstance(v, string_types):
return '"{}"'.format(v)
else:
return v
if filters:
selected_filters = []
for k, v in filters.items():
if isinstance(v, dict):
v = v.keys()[0]
selected_filters.append('row.{} == {}'.format(k, maybe_quote(v)))
code = ' and '.join(selected_filters)
predicate = eval('lambda row: {}'.format(code))
else:
code = None
def x(row):
return True
df = self.analysis.dataframe(predicate, columns=columns, df_class=df_class)
if df is None or df.empty or len(df) == 0:
return None
df['_count'] = 1
def x(x):
return ', '.join(set(str(e) for e in x))
agg = {'_count': 'count'}
for col_name in columns:
c = self.column(col_name)
if c.name == p_dim.name or s_dim and c.name == s_dim.name:
continue
if c.parent and (c.parent == p_dim.name or s_dim and c.parent ==
s_dim.name):
continue
if c.is_measure:
agg[c.name] = np.mean
if c.is_dimension:
agg[c.name] = aggregate_string
plot_axes = [p_dim.name]
if s_dim:
plot_axes.append(s_dim.name)
df = df.groupby(list(columns - set([measure.name]))).agg(agg).reset_index()
df._metadata = ['plot_axes', 'filtered', 'floating', 'labels',
'dimension_set', 'measure']
df.plot_axes = [c for c in plot_axes]
df.filtered = filters
df.floating = list(set(c.name for c in self.primary_dimensions) - set(
df.filtered.keys()) - set(df.plot_axes))
df.labels = [(self.column(c).label.name if self.column(c).label else c) for
c in df.plot_axes]
df.dimension_set = self.dimension_set(p_dim, s_dim=s_dim)
df.measure = measure.name
def x(self):
yield ['id'] + list(df.columns)
for t in df.itertuples():
yield list(t)
df.__class__.rows = property(rows)
return df
| dataframe | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, measure, p_dim, s_dim=None, filters={}, df_class=None):\n import numpy as np\n measure = self.measure(measure)\n p_dim = self.dimension(p_dim)\n assert p_dim\n if s_dim:\n s_dim = self.dimension(s_dim)\n columns = set([measure.name, p_dim.name])\n if p_dim.label:\n if p_dim.geoid:\n columns.add(p_dim.geoid.name)\n columns.add(p_dim.label.name)\n if s_dim:\n columns.add(s_dim.name)\n if s_dim.label:\n columns.add(s_dim.label.name)\n\n def x(v):\n from six import string_types\n if isinstance(v, string_types):\n return '\"{}\"'.format(v)\n else:\n return v\n if filters:\n selected_filters = []\n for k, v in filters.items():\n if isinstance(v, dict):\n v = v.keys()[0]\n selected_filters.append('row.{} == {}'.format(k, maybe_quote(v)))\n code = ' and '.join(selected_filters)\n predicate = eval('lambda row: {}'.format(code))\n else:\n code = None\n\n def x(row):\n return True\n df = self.analysis.dataframe(predicate, columns=columns, df_class=df_class)\n if df is None or df.empty or len(df) == 0:\n return None\n df['_count'] = 1\n\n def x(x):\n return ', '.join(set(str(e) for e in x))\n agg = {'_count': 'count'}\n for col_name in columns:\n c = self.column(col_name)\n if c.name == p_dim.name or s_dim and c.name == s_dim.name:\n continue\n if c.parent and (c.parent == p_dim.name or s_dim and c.parent ==\n s_dim.name):\n continue\n if c.is_measure:\n agg[c.name] = np.mean\n if c.is_dimension:\n agg[c.name] = aggregate_string\n plot_axes = [p_dim.name]\n if s_dim:\n plot_axes.append(s_dim.name)\n df = df.groupby(list(columns - set([measure.name]))).agg(agg).reset_index()\n df._metadata = ['plot_axes', 'filtered', 'floating', 'labels',\n 'dimension_set', 'measure']\n df.plot_axes = [c for c in plot_axes]\n df.filtered = filters\n df.floating = list(set(c.name for c in self.primary_dimensions) - set(\n df.filtered.keys()) - set(df.plot_axes))\n df.labels = [(self.column(c).label.name if self.column(c).label else c) for\n c in df.plot_axes]\n df.dimension_set = self.dimension_set(p_dim, s_dim=s_dim)\n df.measure = measure.name\n\n def x(self):\n yield ['id'] + list(df.columns)\n for t in df.itertuples():\n yield list(t)\n df.__class__.rows = property(rows)\n return df\n",
"role": "user"
},
{
"content": "dataframe",
"role": "assistant"
}
] |
def x(self, p_dim, s_dim=None, dimensions=None, extant=set()):
if not dimensions:
dimensions = self.primary_dimensions
key = p_dim.name
if s_dim:
key += '/' + s_dim.name
if key in extant or p_dim == s_dim:
return
if s_dim and s_dim.valuetype_class.is_geo():
return
extant.add(key)
filtered = {}
for d in dimensions:
if d != p_dim and d != s_dim:
filtered[d.name] = d.pstats.uvalues.keys()
if p_dim.valuetype_class.is_time():
value_type = 'time'
chart_type = 'line'
elif p_dim.valuetype_class.is_geo():
value_type = 'geo'
chart_type = 'map'
else:
value_type = 'general'
chart_type = 'bar'
return dict(key=key, p_dim=p_dim.name, p_dim_type=value_type, p_label=
p_dim.label_or_self.name, s_dim=s_dim.name if s_dim else None,
s_label=s_dim.label_or_self.name if s_dim else None, filters=
filtered, chart_type=chart_type)
| dimension_set | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, p_dim, s_dim=None, dimensions=None, extant=set()):\n if not dimensions:\n dimensions = self.primary_dimensions\n key = p_dim.name\n if s_dim:\n key += '/' + s_dim.name\n if key in extant or p_dim == s_dim:\n return\n if s_dim and s_dim.valuetype_class.is_geo():\n return\n extant.add(key)\n filtered = {}\n for d in dimensions:\n if d != p_dim and d != s_dim:\n filtered[d.name] = d.pstats.uvalues.keys()\n if p_dim.valuetype_class.is_time():\n value_type = 'time'\n chart_type = 'line'\n elif p_dim.valuetype_class.is_geo():\n value_type = 'geo'\n chart_type = 'map'\n else:\n value_type = 'general'\n chart_type = 'bar'\n return dict(key=key, p_dim=p_dim.name, p_dim_type=value_type, p_label=\n p_dim.label_or_self.name, s_dim=s_dim.name if s_dim else None,\n s_label=s_dim.label_or_self.name if s_dim else None, filters=\n filtered, chart_type=chart_type)\n",
"role": "user"
},
{
"content": "dimension_set",
"role": "assistant"
}
] |
def x(self):
for c in self.table.columns:
if c.parent == self.name and 'label' in c.valuetype:
return PartitionColumn(c, self._partition)
| label | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n for c in self.table.columns:\n if c.parent == self.name and 'label' in c.valuetype:\n return PartitionColumn(c, self._partition)\n",
"role": "user"
},
{
"content": "label",
"role": "assistant"
}
] |
def x(self):
from operator import itemgetter
card = self.pstats.nuniques
if self.label:
ig = itemgetter(self.name, self.label.name)
elif self.pstats.nuniques < MAX_LABELS:
ig = itemgetter(self.name, self.name)
else:
return {}
label_set = set()
for row in self._partition:
label_set.add(ig(row))
if len(label_set) >= card:
break
d = dict(label_set)
assert len(d) == len(label_set)
return d
| value_labels | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n from operator import itemgetter\n card = self.pstats.nuniques\n if self.label:\n ig = itemgetter(self.name, self.label.name)\n elif self.pstats.nuniques < MAX_LABELS:\n ig = itemgetter(self.name, self.name)\n else:\n return {}\n label_set = set()\n for row in self._partition:\n label_set.add(ig(row))\n if len(label_set) >= card:\n break\n d = dict(label_set)\n assert len(d) == len(label_set)\n return d\n",
"role": "user"
},
{
"content": "value_labels",
"role": "assistant"
}
] |
def x(page_size=200, page_index=0, sort='', q=''):
data = list_build_configuration_sets_raw(page_size, page_index, sort, q)
if data:
return utils.format_json_list(data)
| list_build_configuration_sets | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(page_size=200, page_index=0, sort='', q=''):\n data = list_build_configuration_sets_raw(page_size, page_index, sort, q)\n if data:\n return utils.format_json_list(data)\n",
"role": "user"
},
{
"content": "list_build_configuration_sets",
"role": "assistant"
}
] |
def x(**kwargs):
config_set = _create_build_config_set_object(**kwargs)
response = utils.checked_api_call(pnc_api.build_group_configs,
'create_new', body=config_set)
if response:
return response.content
| create_build_configuration_set_raw | [
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(**kwargs):\n config_set = _create_build_config_set_object(**kwargs)\n response = utils.checked_api_call(pnc_api.build_group_configs,\n 'create_new', body=config_set)\n if response:\n return response.content\n",
"role": "user"
},
{
"content": "create_build_configuration_set_raw",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.