source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(self, template_name, variables=None): if variables is None: variables = {} template = self._engine.get_template(template_name) return template.render(**variables)
render
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, template_name, variables=None):\n if variables is None:\n variables = {}\n template = self._engine.get_template(template_name)\n return template.render(**variables)\n", "role": "user" }, { "content": "render", "role": "assistant" } ]
def x(self, source, variables=None): if variables is None: variables = {} template = self._engine.from_string(source) return template.render(**variables)
render_source
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, source, variables=None):\n if variables is None:\n variables = {}\n template = self._engine.from_string(source)\n return template.render(**variables)\n", "role": "user" }, { "content": "render_source", "role": "assistant" } ]
def x(url_template, match_whole_str=False, converters=None, default_converter='string', anonymous=False): builder_params = [] url_params = {} result = '^' parts = _split_pattern.split(url_template) for i, part in enumerate(parts): is_url_pattern = _static_url_pattern.match(part) if is_url_pattern: result += re.escape(urlquote(part)) builder_params.append(part) continue is_converter = _converter_pattern.match(part) if is_converter: groups = is_converter.groupdict() converter_name = groups['converter'] or default_converter conv_object = init_converter(converters[converter_name], groups ['args']) variable = groups['variable'] builder_params.append((variable, conv_object)) url_params[variable] = conv_object if anonymous: result += conv_object.regex else: result += '(?P<{}>{})'.format(variable, conv_object.regex) continue raise ValueError('Incorrect url template {!r}'.format(url_template)) if match_whole_str: result += '$' return re.compile(result), url_params, builder_params
construct_re
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url_template, match_whole_str=False, converters=None,\n default_converter='string', anonymous=False):\n builder_params = []\n url_params = {}\n result = '^'\n parts = _split_pattern.split(url_template)\n for i, part in enumerate(parts):\n is_url_pattern = _static_url_pattern.match(part)\n if is_url_pattern:\n result += re.escape(urlquote(part))\n builder_params.append(part)\n continue\n is_converter = _converter_pattern.match(part)\n if is_converter:\n groups = is_converter.groupdict()\n converter_name = groups['converter'] or default_converter\n conv_object = init_converter(converters[converter_name], groups\n ['args'])\n variable = groups['variable']\n builder_params.append((variable, conv_object))\n url_params[variable] = conv_object\n if anonymous:\n result += conv_object.regex\n else:\n result += '(?P<{}>{})'.format(variable, conv_object.regex)\n continue\n raise ValueError('Incorrect url template {!r}'.format(url_template))\n if match_whole_str:\n result += '$'\n return re.compile(result), url_params, builder_params\n", "role": "user" }, { "content": "construct_re", "role": "assistant" } ]
def x(self, path, **kw): m = self._pattern.match(path) if m: kwargs = m.groupdict() for url_arg_name, value_urlencoded in kwargs.items(): conv_obj = self._url_params[url_arg_name] unicode_value = unquote(value_urlencoded) if isinstance(unicode_value, six.binary_type): unicode_value = unicode_value.decode('utf-8', 'replace') try: kwargs[url_arg_name] = conv_obj.to_python(unicode_value, **kw) except ConvertError as err: logger.debug('ConvertError in parameter "%s" by %r, value "%s"' , url_arg_name, err.converter.__class__, err.value) return None, {} return m.group(), kwargs return None, {}
match
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, path, **kw):\n m = self._pattern.match(path)\n if m:\n kwargs = m.groupdict()\n for url_arg_name, value_urlencoded in kwargs.items():\n conv_obj = self._url_params[url_arg_name]\n unicode_value = unquote(value_urlencoded)\n if isinstance(unicode_value, six.binary_type):\n unicode_value = unicode_value.decode('utf-8', 'replace')\n try:\n kwargs[url_arg_name] = conv_obj.to_python(unicode_value, **kw)\n except ConvertError as err:\n logger.debug('ConvertError in parameter \"%s\" by %r, value \"%s\"'\n , url_arg_name, err.converter.__class__, err.value)\n return None, {}\n return m.group(), kwargs\n return None, {}\n", "role": "user" }, { "content": "match", "role": "assistant" } ]
def x(bundle, force=False, force_restricted=False): if not ckan: raise EnvironmentError(MISSING_CREDENTIALS_MSG) try: ckan.action.package_create(**_convert_bundle(bundle)) except ckanapi.ValidationError: if force: logger.warning( '{} dataset already exported, but new export forced. Continue to export dataset stuff.' .format(bundle.dataset)) else: raise access = bundle.dataset.config.metadata.about.access if access == 'restricted' and force_restricted: access = 'private' assert access, 'CKAN publishing requires access level.' if access in ('internal', 'controlled', 'restricted', 'census'): raise UnpublishedAccessError( '{} dataset can not be published because of {} access.'.format( bundle.dataset.vid, bundle.dataset.config.metadata.about.access)) elif access == 'public': user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset. vid.lower(), 'roles': ['editor']}, {'user': 'logged_in', 'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}] elif access == 'registered': user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset. vid.lower(), 'roles': []}, {'user': 'logged_in', 'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}] elif access in ('private', 'licensed', 'test'): user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset. vid.lower(), 'roles': []}, {'user': 'logged_in', 'domain_object': bundle.dataset.vid.lower(), 'roles': []}] organization_users = ckan.action.organization_show(id=CKAN_CONFIG. organization)['users'] for user in organization_users: user_roles.append({'user': user['id'], 'domain_object': bundle. dataset.vid.lower(), 'roles': ['editor']}), for role in user_roles: ckan.action.user_role_update(**role) for partition in bundle.partitions: ckan.action.resource_create(**_convert_partition(partition)) ckan.action.resource_create(**_convert_schema(bundle)) for name, external in six.iteritems(bundle.dataset.config.metadata. external_documentation): ckan.action.resource_create(**_convert_external(bundle, name, external) )
export
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bundle, force=False, force_restricted=False):\n if not ckan:\n raise EnvironmentError(MISSING_CREDENTIALS_MSG)\n try:\n ckan.action.package_create(**_convert_bundle(bundle))\n except ckanapi.ValidationError:\n if force:\n logger.warning(\n '{} dataset already exported, but new export forced. Continue to export dataset stuff.'\n .format(bundle.dataset))\n else:\n raise\n access = bundle.dataset.config.metadata.about.access\n if access == 'restricted' and force_restricted:\n access = 'private'\n assert access, 'CKAN publishing requires access level.'\n if access in ('internal', 'controlled', 'restricted', 'census'):\n raise UnpublishedAccessError(\n '{} dataset can not be published because of {} access.'.format(\n bundle.dataset.vid, bundle.dataset.config.metadata.about.access))\n elif access == 'public':\n user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.\n vid.lower(), 'roles': ['editor']}, {'user': 'logged_in',\n 'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}]\n elif access == 'registered':\n user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.\n vid.lower(), 'roles': []}, {'user': 'logged_in',\n 'domain_object': bundle.dataset.vid.lower(), 'roles': ['editor']}]\n elif access in ('private', 'licensed', 'test'):\n user_roles = [{'user': 'visitor', 'domain_object': bundle.dataset.\n vid.lower(), 'roles': []}, {'user': 'logged_in',\n 'domain_object': bundle.dataset.vid.lower(), 'roles': []}]\n organization_users = ckan.action.organization_show(id=CKAN_CONFIG.\n organization)['users']\n for user in organization_users:\n user_roles.append({'user': user['id'], 'domain_object': bundle.\n dataset.vid.lower(), 'roles': ['editor']}),\n for role in user_roles:\n ckan.action.user_role_update(**role)\n for partition in bundle.partitions:\n ckan.action.resource_create(**_convert_partition(partition))\n ckan.action.resource_create(**_convert_schema(bundle))\n for name, external in six.iteritems(bundle.dataset.config.metadata.\n external_documentation):\n ckan.action.resource_create(**_convert_external(bundle, name, external)\n )\n", "role": "user" }, { "content": "export", "role": "assistant" } ]
def x(bundle): if not ckan: raise EnvironmentError(MISSING_CREDENTIALS_MSG) params = {'q': 'name:{}'.format(bundle.dataset.vid.lower())} resp = ckan.action.package_search(**params) return len(resp['results']) > 0
is_exported
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bundle):\n if not ckan:\n raise EnvironmentError(MISSING_CREDENTIALS_MSG)\n params = {'q': 'name:{}'.format(bundle.dataset.vid.lower())}\n resp = ckan.action.package_search(**params)\n return len(resp['results']) > 0\n", "role": "user" }, { "content": "is_exported", "role": "assistant" } ]
def x(bundle): meta = bundle.dataset.config.metadata notes = '' for f in bundle.dataset.files: if f.path.endswith('documentation.md'): contents = f.unpacked_contents if isinstance(contents, six.binary_type): contents = contents.decode('utf-8') notes = json.dumps(contents) break ret = {'name': bundle.dataset.vid.lower(), 'title': meta.about.title, 'author': meta.contacts.wrangler.name, 'author_email': meta. contacts.wrangler.email, 'maintainer': meta.contacts.maintainer. name, 'maintainer_email': meta.contacts.maintainer.email, 'license_id': '', 'notes': notes, 'url': meta.identity.source, 'version': bundle.dataset.version, 'state': 'active', 'owner_org': CKAN_CONFIG['organization']} return ret
convert_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bundle):\n meta = bundle.dataset.config.metadata\n notes = ''\n for f in bundle.dataset.files:\n if f.path.endswith('documentation.md'):\n contents = f.unpacked_contents\n if isinstance(contents, six.binary_type):\n contents = contents.decode('utf-8')\n notes = json.dumps(contents)\n break\n ret = {'name': bundle.dataset.vid.lower(), 'title': meta.about.title,\n 'author': meta.contacts.wrangler.name, 'author_email': meta.\n contacts.wrangler.email, 'maintainer': meta.contacts.maintainer.\n name, 'maintainer_email': meta.contacts.maintainer.email,\n 'license_id': '', 'notes': notes, 'url': meta.identity.source,\n 'version': bundle.dataset.version, 'state': 'active', 'owner_org':\n CKAN_CONFIG['organization']}\n return ret\n", "role": "user" }, { "content": "convert_bundle", "role": "assistant" } ]
def x(partition): csvfile = six.StringIO() writer = unicodecsv.writer(csvfile) headers = partition.datafile.headers if headers: writer.writerow(headers) for row in partition: writer.writerow([row[h] for h in headers]) csvfile.seek(0) ret = {'package_id': partition.dataset.vid.lower(), 'url': 'http://example.com', 'revision_id': '', 'description': partition. description or '', 'format': 'text/csv', 'hash': '', 'name': partition.name, 'resource_type': '', 'mimetype': 'text/csv', 'mimetype_inner': '', 'webstore_url': '', 'cache_url': '', 'upload': csvfile} return ret
convert_partition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(partition):\n csvfile = six.StringIO()\n writer = unicodecsv.writer(csvfile)\n headers = partition.datafile.headers\n if headers:\n writer.writerow(headers)\n for row in partition:\n writer.writerow([row[h] for h in headers])\n csvfile.seek(0)\n ret = {'package_id': partition.dataset.vid.lower(), 'url':\n 'http://example.com', 'revision_id': '', 'description': partition.\n description or '', 'format': 'text/csv', 'hash': '', 'name':\n partition.name, 'resource_type': '', 'mimetype': 'text/csv',\n 'mimetype_inner': '', 'webstore_url': '', 'cache_url': '', 'upload':\n csvfile}\n return ret\n", "role": "user" }, { "content": "convert_partition", "role": "assistant" } ]
def x(bundle): schema_csv = None for f in bundle.dataset.files: if f.path.endswith('schema.csv'): contents = f.unpacked_contents if isinstance(contents, six.binary_type): contents = contents.decode('utf-8') schema_csv = six.StringIO(contents) schema_csv.seek(0) break ret = {'package_id': bundle.dataset.vid.lower(), 'url': 'http://example.com', 'revision_id': '', 'description': 'Schema of the dataset tables.', 'format': 'text/csv', 'hash': '', 'name': 'schema', 'upload': schema_csv} return ret
convert_schema
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bundle):\n schema_csv = None\n for f in bundle.dataset.files:\n if f.path.endswith('schema.csv'):\n contents = f.unpacked_contents\n if isinstance(contents, six.binary_type):\n contents = contents.decode('utf-8')\n schema_csv = six.StringIO(contents)\n schema_csv.seek(0)\n break\n ret = {'package_id': bundle.dataset.vid.lower(), 'url':\n 'http://example.com', 'revision_id': '', 'description':\n 'Schema of the dataset tables.', 'format': 'text/csv', 'hash': '',\n 'name': 'schema', 'upload': schema_csv}\n return ret\n", "role": "user" }, { "content": "convert_schema", "role": "assistant" } ]
def x(bundle, name, external): ret = {'package_id': bundle.dataset.vid.lower(), 'url': external.url, 'description': external.description, 'name': name} return ret
convert_external
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bundle, name, external):\n ret = {'package_id': bundle.dataset.vid.lower(), 'url': external.url,\n 'description': external.description, 'name': name}\n return ret\n", "role": "user" }, { "content": "convert_external", "role": "assistant" } ]
def x(repository, revision, **kwargs): if not kwargs.get('dependency_ids'): kwargs['dependency_ids'] = [] if not kwargs.get('build_configuration_set_ids'): kwargs['build_configuration_set_ids'] = [] if kwargs.get('generic_parameters'): kwargs['generic_parameters'] = ast.literal_eval(kwargs.get( 'generic_parameters')) if not kwargs.get('project'): kwargs['project'] = pnc_api.projects.get_specific(kwargs.get( 'project_id')).content if not kwargs.get('environment'): kwargs['environment'] = pnc_api.environments.get_specific(kwargs. get('build_environment_id')).content build_configuration = create_build_conf_object(scm_revision=revision, **kwargs) repo_creation = swagger_client.RepositoryCreationUrlAutoRest() repo_creation.scm_url = repository repo_creation.build_configuration_rest = build_configuration response = utils.checked_api_call(pnc_api.bpm, 'start_r_creation_task_with_single_url', body=repo_creation) if response: return response
create_build_configuration_process
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(repository, revision, **kwargs):\n if not kwargs.get('dependency_ids'):\n kwargs['dependency_ids'] = []\n if not kwargs.get('build_configuration_set_ids'):\n kwargs['build_configuration_set_ids'] = []\n if kwargs.get('generic_parameters'):\n kwargs['generic_parameters'] = ast.literal_eval(kwargs.get(\n 'generic_parameters'))\n if not kwargs.get('project'):\n kwargs['project'] = pnc_api.projects.get_specific(kwargs.get(\n 'project_id')).content\n if not kwargs.get('environment'):\n kwargs['environment'] = pnc_api.environments.get_specific(kwargs.\n get('build_environment_id')).content\n build_configuration = create_build_conf_object(scm_revision=revision,\n **kwargs)\n repo_creation = swagger_client.RepositoryCreationUrlAutoRest()\n repo_creation.scm_url = repository\n repo_creation.build_configuration_rest = build_configuration\n response = utils.checked_api_call(pnc_api.bpm,\n 'start_r_creation_task_with_single_url', body=repo_creation)\n if response:\n return response\n", "role": "user" }, { "content": "create_build_configuration_process", "role": "assistant" } ]
def x(repository, no_sync=False): repo = create_repository_configuration_raw(repository, no_sync) if repo: return utils.format_json(repo)
create_repository_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(repository, no_sync=False):\n repo = create_repository_configuration_raw(repository, no_sync)\n if repo:\n return utils.format_json(repo)\n", "role": "user" }, { "content": "create_repository_configuration", "role": "assistant" } ]
def x(task_id, retry=30): success_event_types = 'RC_CREATION_SUCCESS', error_event_types = ('RC_REPO_CREATION_ERROR', 'RC_REPO_CLONE_ERROR', 'RC_CREATION_ERROR') while retry > 0: bpm_task = get_bpm_task_by_id(task_id) if contains_event_type(bpm_task.content.events, success_event_types): break if contains_event_type(bpm_task.content.events, error_event_types): logging.error('Creation of Repository Configuration failed') logging.error(bpm_task.content) return False logging.info( 'Waiting until Repository Configuration creation task ' + str( task_id) + ' finishes.') time.sleep(10) retry -= 1 return retry > 0
wait_for_repo_creation
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(task_id, retry=30):\n success_event_types = 'RC_CREATION_SUCCESS',\n error_event_types = ('RC_REPO_CREATION_ERROR', 'RC_REPO_CLONE_ERROR',\n 'RC_CREATION_ERROR')\n while retry > 0:\n bpm_task = get_bpm_task_by_id(task_id)\n if contains_event_type(bpm_task.content.events, success_event_types):\n break\n if contains_event_type(bpm_task.content.events, error_event_types):\n logging.error('Creation of Repository Configuration failed')\n logging.error(bpm_task.content)\n return False\n logging.info(\n 'Waiting until Repository Configuration creation task ' + str(\n task_id) + ' finishes.')\n time.sleep(10)\n retry -= 1\n return retry > 0\n", "role": "user" }, { "content": "wait_for_repo_creation", "role": "assistant" } ]
def x(self): args = self.format_args if args is None or isinstance(args, dict) and self.count_field not in args: raise TypeError('count is required') return args[self.count_field] if isinstance(args, dict) else args
count
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n args = self.format_args\n if args is None or isinstance(args, dict) and self.count_field not in args:\n raise TypeError('count is required')\n return args[self.count_field] if isinstance(args, dict) else args\n", "role": "user" }, { "content": "count", "role": "assistant" } ]
def x(milestone_id=None, output=None): if not is_input_valid(milestone_id, output): logging.error('invalid input') return 1 create_work_dir(output) download_sources_artifacts(milestone_id, output) create_zip(output)
generate_sources_zip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(milestone_id=None, output=None):\n if not is_input_valid(milestone_id, output):\n logging.error('invalid input')\n return 1\n create_work_dir(output)\n download_sources_artifacts(milestone_id, output)\n create_zip(output)\n", "role": "user" }, { "content": "generate_sources_zip", "role": "assistant" } ]
def x(id): response = utils.checked_api_call(pnc_api.repositories, 'get_specific', id=id) if response: return response.content
get_repository_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(id):\n response = utils.checked_api_call(pnc_api.repositories, 'get_specific',\n id=id)\n if response:\n return response.content\n", "role": "user" }, { "content": "get_repository_configuration", "role": "assistant" } ]
def x(id, external_repository=None, prebuild_sync=None): to_update_id = id rc_to_update = pnc_api.repositories.get_specific(id=to_update_id).content if external_repository is None: external_repository = rc_to_update.external_url else: rc_to_update.external_url = external_repository if prebuild_sync is not None: rc_to_update.pre_build_sync_enabled = prebuild_sync if not external_repository and prebuild_sync: logging.error( 'You cannot enable prebuild sync without external repository') return response = utils.checked_api_call(pnc_api.repositories, 'update', id= to_update_id, body=rc_to_update) if response: return response.content
update_repository_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(id, external_repository=None, prebuild_sync=None):\n to_update_id = id\n rc_to_update = pnc_api.repositories.get_specific(id=to_update_id).content\n if external_repository is None:\n external_repository = rc_to_update.external_url\n else:\n rc_to_update.external_url = external_repository\n if prebuild_sync is not None:\n rc_to_update.pre_build_sync_enabled = prebuild_sync\n if not external_repository and prebuild_sync:\n logging.error(\n 'You cannot enable prebuild sync without external repository')\n return\n response = utils.checked_api_call(pnc_api.repositories, 'update', id=\n to_update_id, body=rc_to_update)\n if response:\n return response.content\n", "role": "user" }, { "content": "update_repository_configuration", "role": "assistant" } ]
def x(page_size=200, page_index=0, sort='', q=''): response = utils.checked_api_call(pnc_api.repositories, 'get_all', page_size=page_size, page_index=page_index, sort=sort, q=q) if response: return utils.format_json_list(response.content)
list_repository_configurations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(page_size=200, page_index=0, sort='', q=''):\n response = utils.checked_api_call(pnc_api.repositories, 'get_all',\n page_size=page_size, page_index=page_index, sort=sort, q=q)\n if response:\n return utils.format_json_list(response.content)\n", "role": "user" }, { "content": "list_repository_configurations", "role": "assistant" } ]
def x(url, page_size=10, page_index=0, sort=''): content = search_repository_configuration_raw(url, page_size, page_index, sort) if content: return utils.format_json_list(content)
search_repository_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url, page_size=10, page_index=0, sort=''):\n content = search_repository_configuration_raw(url, page_size,\n page_index, sort)\n if content:\n return utils.format_json_list(content)\n", "role": "user" }, { "content": "search_repository_configuration", "role": "assistant" } ]
def x(url, page_size=10, page_index=0, sort=''): response = utils.checked_api_call(pnc_api.repositories, 'search', page_size=page_size, page_index=page_index, sort=sort, search=url) if response: return response.content
search_repository_configuration_raw
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url, page_size=10, page_index=0, sort=''):\n response = utils.checked_api_call(pnc_api.repositories, 'search',\n page_size=page_size, page_index=page_index, sort=sort, search=url)\n if response:\n return response.content\n", "role": "user" }, { "content": "search_repository_configuration_raw", "role": "assistant" } ]
def x(url, page_size=10, page_index=0, sort=''): content = match_repository_configuration_raw(url, page_size, page_index, sort) if content: return utils.format_json_list(content)
match_repository_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(url, page_size=10, page_index=0, sort=''):\n content = match_repository_configuration_raw(url, page_size, page_index,\n sort)\n if content:\n return utils.format_json_list(content)\n", "role": "user" }, { "content": "match_repository_configuration", "role": "assistant" } ]
def x(self): return self.env.template.render(self.template, form=self)
render
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return self.env.template.render(self.template, form=self)\n", "role": "user" }, { "content": "render", "role": "assistant" } ]
def x(self, data): self.raw_data = MultiDict(data) self.errors = {} for field in self.fields: if field.writable: self.python_data.update(field.accept()) else: for name in field.field_names: subfield = self.get_field(name) value = self.python_data[subfield.name] subfield.set_raw_value(self.raw_data, subfield.from_python( value)) return self.is_valid
accept
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, data):\n self.raw_data = MultiDict(data)\n self.errors = {}\n for field in self.fields:\n if field.writable:\n self.python_data.update(field.accept())\n else:\n for name in field.field_names:\n subfield = self.get_field(name)\n value = self.python_data[subfield.name]\n subfield.set_raw_value(self.raw_data, subfield.from_python(\n value))\n return self.is_valid\n", "role": "user" }, { "content": "accept", "role": "assistant" } ]
def x(self, compact=True): data = MultiDict() for field in self.fields: raw_value = field.from_python(self.python_data[field.name]) field.set_raw_value(data, raw_value) if compact: data = MultiDict([(k, v) for k, v in data.items() if v]) return data
get_data
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, compact=True):\n data = MultiDict()\n for field in self.fields:\n raw_value = field.from_python(self.python_data[field.name])\n field.set_raw_value(data, raw_value)\n if compact:\n data = MultiDict([(k, v) for k, v in data.items() if v])\n return data\n", "role": "user" }, { "content": "get_data", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.add_configuration_with_http_info(id, **kwargs) else: data = self.add_configuration_with_http_info(id, **kwargs) return data
add_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.add_configuration_with_http_info(id, **kwargs)\n else:\n data = self.add_configuration_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "add_configuration", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.build_with_http_info(id, **kwargs) else: data = self.build_with_http_info(id, **kwargs) return data
build
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.build_with_http_info(id, **kwargs)\n else:\n data = self.build_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "build", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.build_versioned_with_http_info(id, **kwargs) else: data = self.build_versioned_with_http_info(id, **kwargs) return data
build_versioned
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.build_versioned_with_http_info(id, **kwargs)\n else:\n data = self.build_versioned_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "build_versioned", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.delete_specific_with_http_info(id, **kwargs) else: data = self.delete_specific_with_http_info(id, **kwargs) return data
delete_specific
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.delete_specific_with_http_info(id, **kwargs)\n else:\n data = self.delete_specific_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "delete_specific", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.get_all_build_config_set_records_with_http_info(id, ** kwargs) else: data = self.get_all_build_config_set_records_with_http_info(id, ** kwargs) return data
get_all_build_config_set_records
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_all_build_config_set_records_with_http_info(id, **\n kwargs)\n else:\n data = self.get_all_build_config_set_records_with_http_info(id, **\n kwargs)\n return data\n", "role": "user" }, { "content": "get_all_build_config_set_records", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.get_build_records_with_http_info(id, **kwargs) else: data = self.get_build_records_with_http_info(id, **kwargs) return data
get_build_records
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_build_records_with_http_info(id, **kwargs)\n else:\n data = self.get_build_records_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "get_build_records", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.get_configurations_with_http_info(id, **kwargs) else: data = self.get_configurations_with_http_info(id, **kwargs) return data
get_configurations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_configurations_with_http_info(id, **kwargs)\n else:\n data = self.get_configurations_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "get_configurations", "role": "assistant" } ]
def x(self, id, config_id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.remove_configuration_with_http_info(id, config_id, **kwargs ) else: data = self.remove_configuration_with_http_info(id, config_id, **kwargs ) return data
remove_configuration
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, config_id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.remove_configuration_with_http_info(id, config_id, **kwargs\n )\n else:\n data = self.remove_configuration_with_http_info(id, config_id, **kwargs\n )\n return data\n", "role": "user" }, { "content": "remove_configuration", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.update_configurations_with_http_info(id, **kwargs) else: data = self.update_configurations_with_http_info(id, **kwargs) return data
update_configurations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.update_configurations_with_http_info(id, **kwargs)\n else:\n data = self.update_configurations_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "update_configurations", "role": "assistant" } ]
def x(name): package = name.split('.') return [os.path.abspath(os.path.join(x, *package)) for x in sys.path if not os.path.exists(os.path.join(x, *(package + ['__init__.py'])))]
pluginPackagePaths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name):\n package = name.split('.')\n return [os.path.abspath(os.path.join(x, *package)) for x in sys.path if\n not os.path.exists(os.path.join(x, *(package + ['__init__.py'])))]\n", "role": "user" }, { "content": "pluginPackagePaths", "role": "assistant" } ]
def x(func): def x(self, *args, **kwargs): return func(self._root_storage, *args, **kwargs) return wrap
storage_method
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(func):\n\n def x(self, *args, **kwargs):\n return func(self._root_storage, *args, **kwargs)\n return wrap\n", "role": "user" }, { "content": "storage_method", "role": "assistant" } ]
def x(self, template_name, **kw): return self.env.get_template(template_name).render(**kw)
render
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, template_name, **kw):\n return self.env.get_template(template_name).render(**kw)\n", "role": "user" }, { "content": "render", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.cancel_all_builds_in_group_with_http_info(id, **kwargs) else: data = self.cancel_all_builds_in_group_with_http_info(id, **kwargs) return data
cancel_all_builds_in_group
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.cancel_all_builds_in_group_with_http_info(id, **kwargs)\n else:\n data = self.cancel_all_builds_in_group_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "cancel_all_builds_in_group", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.get_all_for_bc_with_http_info(id, **kwargs) else: data = self.get_all_for_bc_with_http_info(id, **kwargs) return data
get_all_for_bc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_all_for_bc_with_http_info(id, **kwargs)\n else:\n data = self.get_all_for_bc_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "get_all_for_bc", "role": "assistant" } ]
def x(self, id, **kwargs): kwargs['_return_http_data_only'] = True if kwargs.get('callback'): return self.get_all_for_bc_set_record_with_http_info(id, **kwargs) else: data = self.get_all_for_bc_set_record_with_http_info(id, **kwargs) return data
get_all_for_bc_set_record
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, id, **kwargs):\n kwargs['_return_http_data_only'] = True\n if kwargs.get('callback'):\n return self.get_all_for_bc_set_record_with_http_info(id, **kwargs)\n else:\n data = self.get_all_for_bc_set_record_with_http_info(id, **kwargs)\n return data\n", "role": "user" }, { "content": "get_all_for_bc_set_record", "role": "assistant" } ]
def x(typename, mapping): def x(self, d, *args, **kwargs): self._original_dict = d def x(path): if not callable(path) and len(path) < 1: raise ValueError( 'Dict paths should be iterables with at least one key or callable objects that take one argument.' ) def x(self): cur_dict = self._original_dict if callable(path): return path(cur_dict) return dict_value_from_path(cur_dict, path) return getter prop_mapping = dict((k, property(getter_from_dict_path(v))) for k, v in mapping.iteritems()) prop_mapping['__init__'] = init return type(typename, tuple(), prop_mapping)
dictmapper
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(typename, mapping):\n\n def x(self, d, *args, **kwargs):\n self._original_dict = d\n\n def x(path):\n if not callable(path) and len(path) < 1:\n raise ValueError(\n 'Dict paths should be iterables with at least one key or callable objects that take one argument.'\n )\n\n def x(self):\n cur_dict = self._original_dict\n if callable(path):\n return path(cur_dict)\n return dict_value_from_path(cur_dict, path)\n return getter\n prop_mapping = dict((k, property(getter_from_dict_path(v))) for k, v in\n mapping.iteritems())\n prop_mapping['__init__'] = init\n return type(typename, tuple(), prop_mapping)\n", "role": "user" }, { "content": "dictmapper", "role": "assistant" } ]
def x(self): return dict(config=self._config, search=self._search, echo=self._echo, read_only=self.read_only)
ctor_args
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n return dict(config=self._config, search=self._search, echo=self._echo,\n read_only=self.read_only)\n", "role": "user" }, { "content": "ctor_args", "role": "assistant" } ]
def x(self, force=False): from ambry.library.config import LibraryConfigSyncProxy lcsp = LibraryConfigSyncProxy(self) lcsp.sync(force=force)
sync_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, force=False):\n from ambry.library.config import LibraryConfigSyncProxy\n lcsp = LibraryConfigSyncProxy(self)\n lcsp.sync(force=force)\n", "role": "user" }, { "content": "sync_config", "role": "assistant" } ]
def x(self): import signal def x(sig, frame): self.log('Trace signal received') self.log(''.join(traceback.format_stack(frame))) signal.signal(signal.SIGUSR2, debug_trace)
init_debug
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n import signal\n\n def x(sig, frame):\n self.log('Trace signal received')\n self.log(''.join(traceback.format_stack(frame)))\n signal.signal(signal.SIGUSR2, debug_trace)\n", "role": "user" }, { "content": "init_debug", "role": "assistant" } ]
def x(self, ref): if not isinstance(ref, ObjectNumber): on = ObjectNumber.parse(ref) else: on = ref ds_on = on.as_dataset return ds_on
resolve_object_number
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref):\n if not isinstance(ref, ObjectNumber):\n on = ObjectNumber.parse(ref)\n else:\n on = ref\n ds_on = on.as_dataset\n return ds_on\n", "role": "user" }, { "content": "resolve_object_number", "role": "assistant" } ]
def x(self, ref, load_all=False, exception=True): return self.database.dataset(ref, load_all=load_all, exception=exception)
dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref, load_all=False, exception=True):\n return self.database.dataset(ref, load_all=load_all, exception=exception)\n", "role": "user" }, { "content": "dataset", "role": "assistant" } ]
def x(self, assignment_class=None, **kwargs): if not ('id' in kwargs and bool(kwargs['id']) ) or assignment_class is not None: kwargs['id'] = self.number(assignment_class) ds = self._db.new_dataset(**kwargs) self._db.commit() b = self.bundle(ds.vid) b.state = Bundle.STATES.NEW b.set_last_access(Bundle.STATES.NEW) b.set_file_system(source_url=self._fs.source(b.identity.source_path), build_url=self._fs.build(b.identity.source_path)) bs_meta = b.build_source_files.file(File.BSFILE.META) bs_meta.set_defaults() bs_meta.record_to_objects() bs_meta.objects_to_record() b.commit() self._db.commit() return b
new_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, assignment_class=None, **kwargs):\n if not ('id' in kwargs and bool(kwargs['id'])\n ) or assignment_class is not None:\n kwargs['id'] = self.number(assignment_class)\n ds = self._db.new_dataset(**kwargs)\n self._db.commit()\n b = self.bundle(ds.vid)\n b.state = Bundle.STATES.NEW\n b.set_last_access(Bundle.STATES.NEW)\n b.set_file_system(source_url=self._fs.source(b.identity.source_path),\n build_url=self._fs.build(b.identity.source_path))\n bs_meta = b.build_source_files.file(File.BSFILE.META)\n bs_meta.set_defaults()\n bs_meta.record_to_objects()\n bs_meta.objects_to_record()\n b.commit()\n self._db.commit()\n return b\n", "role": "user" }, { "content": "new_bundle", "role": "assistant" } ]
def x(self, config): identity = Identity.from_dict(config['identity']) ds = self._db.dataset(identity.vid, exception=False) if not ds: ds = self._db.new_dataset(**identity.dict) b = Bundle(ds, self) b.commit() b.state = Bundle.STATES.NEW b.set_last_access(Bundle.STATES.NEW) return b
new_from_bundle_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, config):\n identity = Identity.from_dict(config['identity'])\n ds = self._db.dataset(identity.vid, exception=False)\n if not ds:\n ds = self._db.new_dataset(**identity.dict)\n b = Bundle(ds, self)\n b.commit()\n b.state = Bundle.STATES.NEW\n b.set_last_access(Bundle.STATES.NEW)\n return b\n", "role": "user" }, { "content": "new_from_bundle_config", "role": "assistant" } ]
def x(self, ref, capture_exceptions=False): from ..orm.exc import NotFoundError if isinstance(ref, Dataset): ds = ref else: try: ds = self._db.dataset(ref) except NotFoundError: ds = None if not ds: try: p = self.partition(ref) ds = p._bundle.dataset except NotFoundError: ds = None if not ds: raise NotFoundError('Failed to find dataset for ref: {}'.format(ref)) b = Bundle(ds, self) b.capture_exceptions = capture_exceptions return b
bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref, capture_exceptions=False):\n from ..orm.exc import NotFoundError\n if isinstance(ref, Dataset):\n ds = ref\n else:\n try:\n ds = self._db.dataset(ref)\n except NotFoundError:\n ds = None\n if not ds:\n try:\n p = self.partition(ref)\n ds = p._bundle.dataset\n except NotFoundError:\n ds = None\n if not ds:\n raise NotFoundError('Failed to find dataset for ref: {}'.format(ref))\n b = Bundle(ds, self)\n b.capture_exceptions = capture_exceptions\n return b\n", "role": "user" }, { "content": "bundle", "role": "assistant" } ]
def x(self, ref, localize=False): if not ref: raise NotFoundError('No partition for empty ref') try: on = ObjectNumber.parse(ref) ds_on = on.as_dataset ds = self._db.dataset(ds_on) self.database.session.refresh(ds) p = ds.partition(ref) except NotObjectNumberError: q = self.database.session.query(Partition).filter(or_(Partition. name == str(ref), Partition.vname == str(ref))).order_by(Partition .vid.desc()) p = q.first() if not p: raise NotFoundError("No partition for ref: '{}'".format(ref)) b = self.bundle(p.d_vid) p = b.wrap_partition(p) if localize: p.localize() return p
partition
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref, localize=False):\n if not ref:\n raise NotFoundError('No partition for empty ref')\n try:\n on = ObjectNumber.parse(ref)\n ds_on = on.as_dataset\n ds = self._db.dataset(ds_on)\n self.database.session.refresh(ds)\n p = ds.partition(ref)\n except NotObjectNumberError:\n q = self.database.session.query(Partition).filter(or_(Partition.\n name == str(ref), Partition.vname == str(ref))).order_by(Partition\n .vid.desc())\n p = q.first()\n if not p:\n raise NotFoundError(\"No partition for ref: '{}'\".format(ref))\n b = self.bundle(p.d_vid)\n p = b.wrap_partition(p)\n if localize:\n p.localize()\n return p\n", "role": "user" }, { "content": "partition", "role": "assistant" } ]
def x(self, ref): try: obj_number = ObjectNumber.parse(ref) ds_obj_number = obj_number.as_dataset dataset = self._db.dataset(ds_obj_number) table = dataset.table(ref) except NotObjectNumberError: q = self.database.session.query(Table).filter(Table.name == str(ref) ).order_by(Table.vid.desc()) table = q.first() if not table: raise NotFoundError("No table for ref: '{}'".format(ref)) return table
table
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref):\n try:\n obj_number = ObjectNumber.parse(ref)\n ds_obj_number = obj_number.as_dataset\n dataset = self._db.dataset(ds_obj_number)\n table = dataset.table(ref)\n except NotObjectNumberError:\n q = self.database.session.query(Table).filter(Table.name == str(ref)\n ).order_by(Table.vid.desc())\n table = q.first()\n if not table:\n raise NotFoundError(\"No table for ref: '{}'\".format(ref))\n return table\n", "role": "user" }, { "content": "table", "role": "assistant" } ]
def x(self, bundle): from six import string_types if isinstance(bundle, string_types): bundle = self.bundle(bundle) self.database.remove_dataset(bundle.dataset)
remove
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, bundle):\n from six import string_types\n if isinstance(bundle, string_types):\n bundle = self.bundle(bundle)\n self.database.remove_dataset(bundle.dataset)\n", "role": "user" }, { "content": "remove", "role": "assistant" } ]
def x(self, b): on = b.identity.on on.revision = on.revision + 1 try: extant = self.bundle(str(on)) if extant: raise ConflictError('Already have a bundle with vid: {}'.format (str(on))) except NotFoundError: pass d = b.dataset.dict d['revision'] = on.revision d['vid'] = str(on) del d['name'] del d['vname'] del d['version'] del d['fqname'] del d['cache_key'] ds = self.database.new_dataset(**d) nb = self.bundle(ds.vid) nb.set_file_system(source_url=b.source_fs.getsyspath('/')) nb.state = Bundle.STATES.NEW nb.commit() for f in b.dataset.files: assert f.major_type == f.MAJOR_TYPE.BUILDSOURCE nb.dataset.files.append(nb.dataset.bsfile(f.minor_type, f.path). update(f)) nb.build_source_files.file(File.BSFILE.META).record_to_objects() nb.build_source_files.file(File.BSFILE.META).objects_to_record() ds.commit() return nb
duplicate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, b):\n on = b.identity.on\n on.revision = on.revision + 1\n try:\n extant = self.bundle(str(on))\n if extant:\n raise ConflictError('Already have a bundle with vid: {}'.format\n (str(on)))\n except NotFoundError:\n pass\n d = b.dataset.dict\n d['revision'] = on.revision\n d['vid'] = str(on)\n del d['name']\n del d['vname']\n del d['version']\n del d['fqname']\n del d['cache_key']\n ds = self.database.new_dataset(**d)\n nb = self.bundle(ds.vid)\n nb.set_file_system(source_url=b.source_fs.getsyspath('/'))\n nb.state = Bundle.STATES.NEW\n nb.commit()\n for f in b.dataset.files:\n assert f.major_type == f.MAJOR_TYPE.BUILDSOURCE\n nb.dataset.files.append(nb.dataset.bsfile(f.minor_type, f.path).\n update(f))\n nb.build_source_files.file(File.BSFILE.META).record_to_objects()\n nb.build_source_files.file(File.BSFILE.META).objects_to_record()\n ds.commit()\n return nb\n", "role": "user" }, { "content": "duplicate", "role": "assistant" } ]
def x(self, db_path, replace=True, cb=None): from ambry.orm.exc import NotFoundError db = Database('sqlite:///{}'.format(db_path)) db.open() if len(db.datasets) == 0: raise NotFoundError('Did not get a dataset in the {} bundle'.format (db_path)) ds = db.dataset(db.datasets[0].vid) assert ds is not None assert ds._database try: b = self.bundle(ds.vid) self.logger.info( "Removing old bundle before checking in new one of same number: '{}'" .format(ds.vid)) self.remove(b) except NotFoundError: pass try: self.dataset(ds.vid) except NotFoundError: self.database.copy_dataset(ds, cb=cb) b = self.bundle(ds.vid) b.commit() self.search.index_bundle(b) return b
checkin_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, db_path, replace=True, cb=None):\n from ambry.orm.exc import NotFoundError\n db = Database('sqlite:///{}'.format(db_path))\n db.open()\n if len(db.datasets) == 0:\n raise NotFoundError('Did not get a dataset in the {} bundle'.format\n (db_path))\n ds = db.dataset(db.datasets[0].vid)\n assert ds is not None\n assert ds._database\n try:\n b = self.bundle(ds.vid)\n self.logger.info(\n \"Removing old bundle before checking in new one of same number: '{}'\"\n .format(ds.vid))\n self.remove(b)\n except NotFoundError:\n pass\n try:\n self.dataset(ds.vid)\n except NotFoundError:\n self.database.copy_dataset(ds, cb=cb)\n b = self.bundle(ds.vid)\n b.commit()\n self.search.index_bundle(b)\n return b\n", "role": "user" }, { "content": "checkin_bundle", "role": "assistant" } ]
def x(self, b, no_partitions=False): raise DeprecationWarning("Don't use any more?") from ambry.bundle.process import call_interval remote_name = self.resolve_remote(b) remote = self.remote(remote_name) db_path = b.package() with b.progress.start('checkin', 0, message='Check in bundle') as ps: ps.add(message='Checking in bundle {} to {}'.format(b.identity. vname, remote)) db_ck = b.identity.cache_key + '.db' ps.add(message='Upload bundle file', item_type='bytes', item_count=0) total = [0] @call_interval(5) def x(n): total[0] += n ps.update(message='Upload bundle file', item_count=total[0]) with open(db_path) as f: remote.makedir(os.path.dirname(db_ck), recursive=True, allow_recreate=True) self.logger.info('Send bundle file {} '.format(db_path)) e = remote.setcontents_async(db_ck, f, progress_callback=upload_cb) e.wait() ps.update(state='done') if not no_partitions: for p in b.partitions: ps.add(message='Upload partition', item_type='bytes', item_count=0, p_vid=p.vid) with p.datafile.open(mode='rb') as fin: total = [0] @call_interval(5) def x(bytes): total[0] += bytes ps.update(message='Upload partition'.format(p. identity.vname), item_count=total[0]) remote.makedir(os.path.dirname(p.datafile.path), recursive=True, allow_recreate=True) event = remote.setcontents_async(p.datafile.path, fin, progress_callback=progress) event.wait() ps.update(state='done') ps.add(message='Setting metadata') ident = json.dumps(b.identity.dict) remote.setcontents(os.path.join('_meta', 'vid', b.identity.vid), ident) remote.setcontents(os.path.join('_meta', 'id', b.identity.id_), ident) remote.setcontents(os.path.join('_meta', 'vname', text_type(b. identity.vname)), ident) remote.setcontents(os.path.join('_meta', 'name', text_type(b. identity.name)), ident) ps.update(state='done') b.dataset.commit() return remote_name, db_ck
send_to_remote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, b, no_partitions=False):\n raise DeprecationWarning(\"Don't use any more?\")\n from ambry.bundle.process import call_interval\n remote_name = self.resolve_remote(b)\n remote = self.remote(remote_name)\n db_path = b.package()\n with b.progress.start('checkin', 0, message='Check in bundle') as ps:\n ps.add(message='Checking in bundle {} to {}'.format(b.identity.\n vname, remote))\n db_ck = b.identity.cache_key + '.db'\n ps.add(message='Upload bundle file', item_type='bytes', item_count=0)\n total = [0]\n\n @call_interval(5)\n def x(n):\n total[0] += n\n ps.update(message='Upload bundle file', item_count=total[0])\n with open(db_path) as f:\n remote.makedir(os.path.dirname(db_ck), recursive=True,\n allow_recreate=True)\n self.logger.info('Send bundle file {} '.format(db_path))\n e = remote.setcontents_async(db_ck, f, progress_callback=upload_cb)\n e.wait()\n ps.update(state='done')\n if not no_partitions:\n for p in b.partitions:\n ps.add(message='Upload partition', item_type='bytes',\n item_count=0, p_vid=p.vid)\n with p.datafile.open(mode='rb') as fin:\n total = [0]\n\n @call_interval(5)\n def x(bytes):\n total[0] += bytes\n ps.update(message='Upload partition'.format(p.\n identity.vname), item_count=total[0])\n remote.makedir(os.path.dirname(p.datafile.path),\n recursive=True, allow_recreate=True)\n event = remote.setcontents_async(p.datafile.path, fin,\n progress_callback=progress)\n event.wait()\n ps.update(state='done')\n ps.add(message='Setting metadata')\n ident = json.dumps(b.identity.dict)\n remote.setcontents(os.path.join('_meta', 'vid', b.identity.vid), ident)\n remote.setcontents(os.path.join('_meta', 'id', b.identity.id_), ident)\n remote.setcontents(os.path.join('_meta', 'vname', text_type(b.\n identity.vname)), ident)\n remote.setcontents(os.path.join('_meta', 'name', text_type(b.\n identity.name)), ident)\n ps.update(state='done')\n b.dataset.commit()\n return remote_name, db_ck\n", "role": "user" }, { "content": "send_to_remote", "role": "assistant" } ]
def x(self, ref, remote=None): if not remote: remote, vname = self.find_remote_bundle(ref) if vname: ref = vname else: pass if not remote: raise NotFoundError("Failed to find bundle ref '{}' in any remote". format(ref)) self.logger.info("Load '{}' from '{}'".format(ref, remote)) vid = self._checkin_remote_bundle(remote, ref) self.commit() return vid
checkin_remote_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref, remote=None):\n if not remote:\n remote, vname = self.find_remote_bundle(ref)\n if vname:\n ref = vname\n else:\n pass\n if not remote:\n raise NotFoundError(\"Failed to find bundle ref '{}' in any remote\".\n format(ref))\n self.logger.info(\"Load '{}' from '{}'\".format(ref, remote))\n vid = self._checkin_remote_bundle(remote, ref)\n self.commit()\n return vid\n", "role": "user" }, { "content": "checkin_remote_bundle", "role": "assistant" } ]
def x(self, remote, ref): from ambry.bundle.process import call_interval from ambry.orm.exc import NotFoundError from ambry.orm import Remote from ambry.util.flo import copy_file_or_flo from tempfile import NamedTemporaryFile assert isinstance(remote, Remote) @call_interval(5) def x(r, total): self.logger.info('{}: Downloaded {} bytes'.format(ref, total)) b = None try: b = self.bundle(ref) self.logger.info('{}: Already installed'.format(ref)) vid = b.identity.vid except NotFoundError: self.logger.info('{}: Syncing'.format(ref)) db_dir = self.filesystem.downloads('bundles') db_f = os.path.join(db_dir, ref) if not os.path.exists(os.path.join(db_dir, db_f)): self.logger.info("Downloading bundle '{}' to '{}".format(ref, db_f) ) with open(db_f, 'wb') as f_out: with remote.checkout(ref) as f: copy_file_or_flo(f, f_out, cb=cb) f_out.flush() self.checkin_bundle(db_f) b = self.bundle(ref) b.dataset.data['remote_name'] = remote.short_name b.dataset.upstream = remote.url b.dstate = b.STATES.CHECKEDOUT b.commit() finally: if b: b.progress.close() vid = b.identity.vid return vid
checkin_remote_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, remote, ref):\n from ambry.bundle.process import call_interval\n from ambry.orm.exc import NotFoundError\n from ambry.orm import Remote\n from ambry.util.flo import copy_file_or_flo\n from tempfile import NamedTemporaryFile\n assert isinstance(remote, Remote)\n\n @call_interval(5)\n def x(r, total):\n self.logger.info('{}: Downloaded {} bytes'.format(ref, total))\n b = None\n try:\n b = self.bundle(ref)\n self.logger.info('{}: Already installed'.format(ref))\n vid = b.identity.vid\n except NotFoundError:\n self.logger.info('{}: Syncing'.format(ref))\n db_dir = self.filesystem.downloads('bundles')\n db_f = os.path.join(db_dir, ref)\n if not os.path.exists(os.path.join(db_dir, db_f)):\n self.logger.info(\"Downloading bundle '{}' to '{}\".format(ref, db_f)\n )\n with open(db_f, 'wb') as f_out:\n with remote.checkout(ref) as f:\n copy_file_or_flo(f, f_out, cb=cb)\n f_out.flush()\n self.checkin_bundle(db_f)\n b = self.bundle(ref)\n b.dataset.data['remote_name'] = remote.short_name\n b.dataset.upstream = remote.url\n b.dstate = b.STATES.CHECKEDOUT\n b.commit()\n finally:\n if b:\n b.progress.close()\n vid = b.identity.vid\n return vid\n", "role": "user" }, { "content": "checkin_remote_bundle", "role": "assistant" } ]
def x(self): from ambry.orm import Remote for r in self.database.session.query(Remote).all(): if not r.short_name: continue yield self.remote(r.short_name)
remotes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.orm import Remote\n for r in self.database.session.query(Remote).all():\n if not r.short_name:\n continue\n yield self.remote(r.short_name)\n", "role": "user" }, { "content": "remotes", "role": "assistant" } ]
def x(self, name): from ambry.orm import Remote from sqlalchemy import or_ from ambry.orm.exc import NotFoundError from sqlalchemy.orm.exc import NoResultFound, MultipleResultsFound if not name.strip(): raise NotFoundError('Empty remote name') try: try: r = self.database.session.query(Remote).filter(Remote. short_name == name).one() except NoResultFound as e: r = None if not r: r = self.database.session.query(Remote).filter(Remote.url == name ).one() except NoResultFound as e: raise NotFoundError(str(e) + '; ' + name) except MultipleResultsFound as e: self.logger.error("Got multiple results for search for remote '{}': {}" .format(name, e)) return None return r
remote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, name):\n from ambry.orm import Remote\n from sqlalchemy import or_\n from ambry.orm.exc import NotFoundError\n from sqlalchemy.orm.exc import NoResultFound, MultipleResultsFound\n if not name.strip():\n raise NotFoundError('Empty remote name')\n try:\n try:\n r = self.database.session.query(Remote).filter(Remote.\n short_name == name).one()\n except NoResultFound as e:\n r = None\n if not r:\n r = self.database.session.query(Remote).filter(Remote.url == name\n ).one()\n except NoResultFound as e:\n raise NotFoundError(str(e) + '; ' + name)\n except MultipleResultsFound as e:\n self.logger.error(\"Got multiple results for search for remote '{}': {}\"\n .format(name, e))\n return None\n return r\n", "role": "user" }, { "content": "remote", "role": "assistant" } ]
def x(self, ref, remote_service_type='s3'): for r in self.remotes: if remote_service_type and r.service != remote_service_type: continue if 'list' not in r.data: continue for k, v in r.data['list'].items(): if ref in v.values(): return r, v['vname'] return None, None
find_remote_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref, remote_service_type='s3'):\n for r in self.remotes:\n if remote_service_type and r.service != remote_service_type:\n continue\n if 'list' not in r.data:\n continue\n for k, v in r.data['list'].items():\n if ref in v.values():\n return r, v['vname']\n return None, None\n", "role": "user" }, { "content": "find_remote_bundle", "role": "assistant" } ]
def x(self, ref, try_harder=None): from ambry.identity import ObjectNumber remote, vid = self._find_remote_bundle(ref) if remote: return remote, vid if try_harder: on = ObjectNumber.parse(vid) if on: raise NotImplementedError() don = on.as_dataset return self._find_remote_bundle(vid) parts = ref.split('-') for i in range(len(parts) - 1, 2, -1): remote, vid = self._find_remote_bundle('-'.join(parts[:i])) if remote: return remote, vid return None, None
find_remote_bundle
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ref, try_harder=None):\n from ambry.identity import ObjectNumber\n remote, vid = self._find_remote_bundle(ref)\n if remote:\n return remote, vid\n if try_harder:\n on = ObjectNumber.parse(vid)\n if on:\n raise NotImplementedError()\n don = on.as_dataset\n return self._find_remote_bundle(vid)\n parts = ref.split('-')\n for i in range(len(parts) - 1, 2, -1):\n remote, vid = self._find_remote_bundle('-'.join(parts[:i]))\n if remote:\n return remote, vid\n return None, None\n", "role": "user" }, { "content": "find_remote_bundle", "role": "assistant" } ]
def x(self, url): from sqlalchemy.orm.exc import NoResultFound from ambry.orm.exc import NotFoundError from ambry.util import parse_url_to_dict from ambry.orm import Account pd = parse_url_to_dict(url) try: act = self.database.session.query(Account).filter(Account. account_id == pd['netloc']).one() act.secret_password = self._account_password return act except NoResultFound: pass for r in self.remotes: if url.startswith(r.url): return r raise NotFoundError("Did not find account for url: '{}' ".format(url))
account
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, url):\n from sqlalchemy.orm.exc import NoResultFound\n from ambry.orm.exc import NotFoundError\n from ambry.util import parse_url_to_dict\n from ambry.orm import Account\n pd = parse_url_to_dict(url)\n try:\n act = self.database.session.query(Account).filter(Account.\n account_id == pd['netloc']).one()\n act.secret_password = self._account_password\n return act\n except NoResultFound:\n pass\n for r in self.remotes:\n if url.startswith(r.url):\n return r\n raise NotFoundError(\"Did not find account for url: '{}' \".format(url))\n", "role": "user" }, { "content": "account", "role": "assistant" } ]
def x(self): d = {} if False and not self._account_password: from ambry.dbexceptions import ConfigurationError raise ConfigurationError( "Can't access accounts without setting an account password either in the accounts.password config, or in the AMBRY_ACCOUNT_PASSWORD env var." ) for act in self.database.session.query(Account).all(): if self._account_password: act.secret_password = self._account_password e = act.dict a_id = e['account_id'] d[a_id] = e return d
accounts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n d = {}\n if False and not self._account_password:\n from ambry.dbexceptions import ConfigurationError\n raise ConfigurationError(\n \"Can't access accounts without setting an account password either in the accounts.password config, or in the AMBRY_ACCOUNT_PASSWORD env var.\"\n )\n for act in self.database.session.query(Account).all():\n if self._account_password:\n act.secret_password = self._account_password\n e = act.dict\n a_id = e['account_id']\n d[a_id] = e\n return d\n", "role": "user" }, { "content": "accounts", "role": "assistant" } ]
def x(self, assignment_class=None, namespace='d'): if assignment_class == 'self': return str(DatasetNumber()) elif assignment_class is None: try: nsconfig = self.services['numbers'] except ConfigurationError: self.logger.error( 'No number server configuration; returning self assigned number' ) return str(DatasetNumber()) for assignment_class in ('self', 'unregistered', 'registered', 'authority'): if assignment_class + '-key' in nsconfig: break if assignment_class == 'self': return str(DatasetNumber()) else: try: nsconfig = self.services['numbers'] except ConfigurationError: raise ConfigurationError('No number server configuration') if assignment_class + '-key' not in nsconfig: raise ConfigurationError( 'Assignment class {} not number server config'.format( assignment_class)) try: key = nsconfig[assignment_class + '-key'] config = {'key': key, 'host': nsconfig['host'], 'port': nsconfig. get('port', 80)} ns = NumberServer(**config) n = str(next(ns)) self.logger.info('Got number from number server: {}'.format(n)) except HTTPError as e: self.logger.error('Failed to get number from number server for key: {}' .format(key, e.message)) self.logger.error( 'Using self-generated number. There is no problem with this, but they are longer than centrally generated numbers.' ) n = str(DatasetNumber()) return n
number
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, assignment_class=None, namespace='d'):\n if assignment_class == 'self':\n return str(DatasetNumber())\n elif assignment_class is None:\n try:\n nsconfig = self.services['numbers']\n except ConfigurationError:\n self.logger.error(\n 'No number server configuration; returning self assigned number'\n )\n return str(DatasetNumber())\n for assignment_class in ('self', 'unregistered', 'registered',\n 'authority'):\n if assignment_class + '-key' in nsconfig:\n break\n if assignment_class == 'self':\n return str(DatasetNumber())\n else:\n try:\n nsconfig = self.services['numbers']\n except ConfigurationError:\n raise ConfigurationError('No number server configuration')\n if assignment_class + '-key' not in nsconfig:\n raise ConfigurationError(\n 'Assignment class {} not number server config'.format(\n assignment_class))\n try:\n key = nsconfig[assignment_class + '-key']\n config = {'key': key, 'host': nsconfig['host'], 'port': nsconfig.\n get('port', 80)}\n ns = NumberServer(**config)\n n = str(next(ns))\n self.logger.info('Got number from number server: {}'.format(n))\n except HTTPError as e:\n self.logger.error('Failed to get number from number server for key: {}'\n .format(key, e.message))\n self.logger.error(\n 'Using self-generated number. There is no problem with this, but they are longer than centrally generated numbers.'\n )\n n = str(DatasetNumber())\n return n\n", "role": "user" }, { "content": "number", "role": "assistant" } ]
def x(self): ret = self._db.session.query(Config).filter(Config.type == 'buildstate' ).filter(Config.group == 'access').filter(Config.key == 'last' ).order_by(Config.modified.desc()).all() return ret
edit_history
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n ret = self._db.session.query(Config).filter(Config.type == 'buildstate'\n ).filter(Config.group == 'access').filter(Config.key == 'last'\n ).order_by(Config.modified.desc()).all()\n return ret\n", "role": "user" }, { "content": "edit_history", "role": "assistant" } ]
def x(self, dir, detach=False, force=False): import yaml fs = fsopendir(dir) bundles = [] for f in fs.walkfiles(wildcard='bundle.yaml'): self.logger.info('Visiting {}'.format(f)) config = yaml.load(fs.getcontents(f)) if not config: self.logger.error( "Failed to get a valid bundle configuration from '{}'". format(f)) bid = config['identity']['id'] try: b = self.bundle(bid) except NotFoundError: b = None if not b: b = self.new_from_bundle_config(config) self.logger.info('{} Loading New'.format(b.identity.fqname)) else: self.logger.info('{} Loading Existing'.format(b.identity.fqname)) source_url = os.path.dirname(fs.getsyspath(f)) b.set_file_system(source_url=source_url) self.logger.info('{} Loading from {}'.format(b.identity.fqname, source_url)) b.sync_in() if detach: self.logger.info('{} Detaching'.format(b.identity.fqname)) b.set_file_system(source_url=None) if force: self.logger.info('{} Sync out'.format(b.identity.fqname)) b = self.bundle(bid) b.sync_out() bundles.append(b) b.close() return bundles
import_bundles
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dir, detach=False, force=False):\n import yaml\n fs = fsopendir(dir)\n bundles = []\n for f in fs.walkfiles(wildcard='bundle.yaml'):\n self.logger.info('Visiting {}'.format(f))\n config = yaml.load(fs.getcontents(f))\n if not config:\n self.logger.error(\n \"Failed to get a valid bundle configuration from '{}'\".\n format(f))\n bid = config['identity']['id']\n try:\n b = self.bundle(bid)\n except NotFoundError:\n b = None\n if not b:\n b = self.new_from_bundle_config(config)\n self.logger.info('{} Loading New'.format(b.identity.fqname))\n else:\n self.logger.info('{} Loading Existing'.format(b.identity.fqname))\n source_url = os.path.dirname(fs.getsyspath(f))\n b.set_file_system(source_url=source_url)\n self.logger.info('{} Loading from {}'.format(b.identity.fqname,\n source_url))\n b.sync_in()\n if detach:\n self.logger.info('{} Detaching'.format(b.identity.fqname))\n b.set_file_system(source_url=None)\n if force:\n self.logger.info('{} Sync out'.format(b.identity.fqname))\n b = self.bundle(bid)\n b.sync_out()\n bundles.append(b)\n b.close()\n return bundles\n", "role": "user" }, { "content": "import_bundles", "role": "assistant" } ]
def x(self, limited_run=False): from multiprocessing import cpu_count from ambry.bundle.concurrent import Pool, init_library if self.processes: cpus = self.processes else: cpus = cpu_count() self.logger.info('Starting MP pool with {} processors'.format(cpus)) return Pool(self, processes=cpus, initializer=init_library, maxtasksperchild=1, initargs=[self.database.dsn, self. _account_password, limited_run])
process_pool
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, limited_run=False):\n from multiprocessing import cpu_count\n from ambry.bundle.concurrent import Pool, init_library\n if self.processes:\n cpus = self.processes\n else:\n cpus = cpu_count()\n self.logger.info('Starting MP pool with {} processors'.format(cpus))\n return Pool(self, processes=cpus, initializer=init_library,\n maxtasksperchild=1, initargs=[self.database.dsn, self.\n _account_password, limited_run])\n", "role": "user" }, { "content": "process_pool", "role": "assistant" } ]
def x(): import sys import inspect try: raise Exception except: file_ = '.../' + '/'.join(inspect.currentframe().f_code.co_filename .split('/')[-3:]) line_ = sys.exc_info()[2].tb_frame.f_back.f_lineno return '{}:{}'.format(file_, line_)
file_loc
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n import sys\n import inspect\n try:\n raise Exception\n except:\n file_ = '.../' + '/'.join(inspect.currentframe().f_code.co_filename\n .split('/')[-3:])\n line_ = sys.exc_info()[2].tb_frame.f_back.f_lineno\n return '{}:{}'.format(file_, line_)\n", "role": "user" }, { "content": "file_loc", "role": "assistant" } ]
def x(bundle, source_headers, dest_table, env): dest_headers = [c.name for c in dest_table.columns] row_processors = [] out = [file_header] transforms = list(dest_table.transforms) column_names = [] column_types = [] for i, segments in enumerate(transforms): seg_funcs = [] for col_num, (segment, column) in enumerate(zip(segments, dest_table.columns), 1): if not segment: seg_funcs.append('row[{}]'.format(col_num - 1)) continue assert column assert column.name == segment['column'].name col_name = column.name preamble, try_lines, exception = make_stack(env, i, segment) assert col_num == column.sequence_id, (dest_table.name, col_num, column.sequence_id) column_names.append(col_name) column_types.append(column.datatype) f_name = '{table_name}_{column_name}_{stage}'.format(table_name =dest_table.name, column_name=col_name, stage=i) exception = (exception if exception else 'raise ValueError("Failed to cast column \'{}\', in function {}, value \'{}\': {}".format(header_d,"' + f_name + '", v.encode(\'ascii\', \'replace\'), exc) ) ') try: i_s = source_headers.index(column.name) header_s = column.name v = 'row[{}]'.format(i_s) except ValueError as e: i_s = 'None' header_s = None v = 'None' if col_num > 1 else 'row_n' i_d = column.sequence_id - 1 header_d = column.name template_args = dict(f_name=f_name, table_name=dest_table.name, column_name=col_name, stage=i, i_s=i_s, i_d=i_d, header_s= header_s, header_d=header_d, v=v, exception=indent + exception, stack='\n'.join(indent + l for l in try_lines), col_args='# col_args not implemented yet') seg_funcs.append(f_name + "({v}, {i_s}, {i_d}, {header_s}, '{header_d}', row, row_n, errors, scratch, accumulator, pipe, bundle, source)" .format(v=v, i_s=i_s, i_d=i_d, header_s="'" + header_s + "'" if header_s else 'None', header_d=header_d)) out.append('\n'.join(preamble)) out.append(column_template.format(**template_args)) source_headers = dest_headers stack = '\n'.join('{}{}, # {}'.format(indent, l, cn) for l, cn, dt in zip(seg_funcs, column_names, column_types)) out.append(row_template.format(table=dest_table.name, stage=i, stack=stack)) row_processors.append('row_{table}_{stage}'.format(stage=i, table= dest_table.name)) stack = '\n'.join("{}cast_{}(row[{}], '{}', errors),".format(indent, c. datatype, i, c.name) for i, c in enumerate(dest_table.columns)) out.append(row_template.format(table=dest_table.name, stage=len( transforms), stack=stack)) row_processors.append('row_{table}_{stage}'.format(stage=len(transforms ), table=dest_table.name)) out.append('row_processors = [{}]'.format(','.join(row_processors))) return '\n'.join(out)
make_row_processors
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(bundle, source_headers, dest_table, env):\n dest_headers = [c.name for c in dest_table.columns]\n row_processors = []\n out = [file_header]\n transforms = list(dest_table.transforms)\n column_names = []\n column_types = []\n for i, segments in enumerate(transforms):\n seg_funcs = []\n for col_num, (segment, column) in enumerate(zip(segments,\n dest_table.columns), 1):\n if not segment:\n seg_funcs.append('row[{}]'.format(col_num - 1))\n continue\n assert column\n assert column.name == segment['column'].name\n col_name = column.name\n preamble, try_lines, exception = make_stack(env, i, segment)\n assert col_num == column.sequence_id, (dest_table.name, col_num,\n column.sequence_id)\n column_names.append(col_name)\n column_types.append(column.datatype)\n f_name = '{table_name}_{column_name}_{stage}'.format(table_name\n =dest_table.name, column_name=col_name, stage=i)\n exception = (exception if exception else \n 'raise ValueError(\"Failed to cast column \\'{}\\', in function {}, value \\'{}\\': {}\".format(header_d,\"'\n + f_name + '\", v.encode(\\'ascii\\', \\'replace\\'), exc) ) ')\n try:\n i_s = source_headers.index(column.name)\n header_s = column.name\n v = 'row[{}]'.format(i_s)\n except ValueError as e:\n i_s = 'None'\n header_s = None\n v = 'None' if col_num > 1 else 'row_n'\n i_d = column.sequence_id - 1\n header_d = column.name\n template_args = dict(f_name=f_name, table_name=dest_table.name,\n column_name=col_name, stage=i, i_s=i_s, i_d=i_d, header_s=\n header_s, header_d=header_d, v=v, exception=indent +\n exception, stack='\\n'.join(indent + l for l in try_lines),\n col_args='# col_args not implemented yet')\n seg_funcs.append(f_name +\n \"({v}, {i_s}, {i_d}, {header_s}, '{header_d}', row, row_n, errors, scratch, accumulator, pipe, bundle, source)\"\n .format(v=v, i_s=i_s, i_d=i_d, header_s=\"'\" + header_s +\n \"'\" if header_s else 'None', header_d=header_d))\n out.append('\\n'.join(preamble))\n out.append(column_template.format(**template_args))\n source_headers = dest_headers\n stack = '\\n'.join('{}{}, # {}'.format(indent, l, cn) for l, cn, dt in\n zip(seg_funcs, column_names, column_types))\n out.append(row_template.format(table=dest_table.name, stage=i,\n stack=stack))\n row_processors.append('row_{table}_{stage}'.format(stage=i, table=\n dest_table.name))\n stack = '\\n'.join(\"{}cast_{}(row[{}], '{}', errors),\".format(indent, c.\n datatype, i, c.name) for i, c in enumerate(dest_table.columns))\n out.append(row_template.format(table=dest_table.name, stage=len(\n transforms), stack=stack))\n row_processors.append('row_{table}_{stage}'.format(stage=len(transforms\n ), table=dest_table.name))\n out.append('row_processors = [{}]'.format(','.join(row_processors)))\n return '\\n'.join(out)\n", "role": "user" }, { "content": "make_row_processors", "role": "assistant" } ]
def x(f, f_name=None, raise_for_missing=True): import inspect from ambry.dbexceptions import ConfigurationError if inspect.isclass(f): try: args = inspect.getargspec(f.__init__).args except TypeError as e: raise TypeError('Failed to inspect {}: {}'.format(f, e)) else: args = inspect.getargspec(f).args if len(args) > 1 and args[0] == 'self': args = args[1:] for a in args: if a not in all_args + ('exception',): if raise_for_missing: raise ConfigurationError( "Caster code {} has unknown argument name: '{}'. Must be one of: {} " .format(f, a, ','.join(all_args))) arg_map = {e: e for e in var_args} args = [arg_map.get(a, a) for a in args] return '{}({})'.format(f_name if f_name else f.__name__, ','.join(args))
calling_code
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(f, f_name=None, raise_for_missing=True):\n import inspect\n from ambry.dbexceptions import ConfigurationError\n if inspect.isclass(f):\n try:\n args = inspect.getargspec(f.__init__).args\n except TypeError as e:\n raise TypeError('Failed to inspect {}: {}'.format(f, e))\n else:\n args = inspect.getargspec(f).args\n if len(args) > 1 and args[0] == 'self':\n args = args[1:]\n for a in args:\n if a not in all_args + ('exception',):\n if raise_for_missing:\n raise ConfigurationError(\n \"Caster code {} has unknown argument name: '{}'. Must be one of: {} \"\n .format(f, a, ','.join(all_args)))\n arg_map = {e: e for e in var_args}\n args = [arg_map.get(a, a) for a in args]\n return '{}({})'.format(f_name if f_name else f.__name__, ','.join(args))\n", "role": "user" }, { "content": "calling_code", "role": "assistant" } ]
def x(env, stage, segment): import string import random from ambry.valuetype import ValueType column = segment['column'] def x(column, t): preamble = [] line_t = 'v = {} # {}' if isinstance(t, type) and issubclass(t, ValueType): try: cc, fl = calling_code(t, t.__name__), file_loc() except TypeError: cc, fl = '{}(v)'.format(t.__name__), file_loc() preamble.append("{} = resolve_value_type('{}') # {}".format(t. __name__, t.vt_code, fl)) elif isinstance(t, type): cc, fl = 'parse_{}(v, header_d)'.format(t.__name__), file_loc() elif callable(env.get(t)): cc, fl = calling_code(env.get(t), t), file_loc() else: rnd = ''.join(random.choice(string.ascii_lowercase) for _ in range(6)) name = 'tg_{}_{}_{}'.format(column.name, stage, rnd) try: a, b, fl = rewrite_tg(env, name, t) except CodeGenError as e: raise CodeGenError( "Failed to re-write pipe code '{}' in column '{}.{}': {} " .format(t, column.table.name, column.name, e)) cc = str(a) if b: preamble.append('{} = {} # {}'.format(name, b, fl)) line = line_t.format(cc, fl) return line, preamble preamble = [] try_lines = [] for t in ([segment['init'], segment['datatype']] + segment['transforms']): if not t: continue line, col_preamble = make_line(column, t) preamble += col_preamble try_lines.append(line) exception = None if segment['exception']: exception, col_preamble = make_line(column, segment['exception']) if len(try_lines) == 0: try_lines.append('pass # Empty pipe segment') assert len(try_lines) > 0, column.name return preamble, try_lines, exception
make_stack
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(env, stage, segment):\n import string\n import random\n from ambry.valuetype import ValueType\n column = segment['column']\n\n def x(column, t):\n preamble = []\n line_t = 'v = {} # {}'\n if isinstance(t, type) and issubclass(t, ValueType):\n try:\n cc, fl = calling_code(t, t.__name__), file_loc()\n except TypeError:\n cc, fl = '{}(v)'.format(t.__name__), file_loc()\n preamble.append(\"{} = resolve_value_type('{}') # {}\".format(t.\n __name__, t.vt_code, fl))\n elif isinstance(t, type):\n cc, fl = 'parse_{}(v, header_d)'.format(t.__name__), file_loc()\n elif callable(env.get(t)):\n cc, fl = calling_code(env.get(t), t), file_loc()\n else:\n rnd = ''.join(random.choice(string.ascii_lowercase) for _ in\n range(6))\n name = 'tg_{}_{}_{}'.format(column.name, stage, rnd)\n try:\n a, b, fl = rewrite_tg(env, name, t)\n except CodeGenError as e:\n raise CodeGenError(\n \"Failed to re-write pipe code '{}' in column '{}.{}': {} \"\n .format(t, column.table.name, column.name, e))\n cc = str(a)\n if b:\n preamble.append('{} = {} # {}'.format(name, b, fl))\n line = line_t.format(cc, fl)\n return line, preamble\n preamble = []\n try_lines = []\n for t in ([segment['init'], segment['datatype']] + segment['transforms']):\n if not t:\n continue\n line, col_preamble = make_line(column, t)\n preamble += col_preamble\n try_lines.append(line)\n exception = None\n if segment['exception']:\n exception, col_preamble = make_line(column, segment['exception'])\n if len(try_lines) == 0:\n try_lines.append('pass # Empty pipe segment')\n assert len(try_lines) > 0, column.name\n return preamble, try_lines, exception\n", "role": "user" }, { "content": "make_stack", "role": "assistant" } ]
def x(env, tg_name, code): visitor = ReplaceTG(env, tg_name) assert visitor.tg_name tree = visitor.visit(ast.parse(code)) if visitor.loc: loc = ' #' + visitor.loc else: loc = file_loc() if visitor.trans_gen: tg = meta.dump_python_source(visitor.trans_gen).strip() else: tg = None return meta.dump_python_source(tree).strip(), tg, loc
rewrite_tg
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(env, tg_name, code):\n visitor = ReplaceTG(env, tg_name)\n assert visitor.tg_name\n tree = visitor.visit(ast.parse(code))\n if visitor.loc:\n loc = ' #' + visitor.loc\n else:\n loc = file_loc()\n if visitor.trans_gen:\n tg = meta.dump_python_source(visitor.trans_gen).strip()\n else:\n tg = None\n return meta.dump_python_source(tree).strip(), tg, loc\n", "role": "user" }, { "content": "rewrite_tg", "role": "assistant" } ]
def x(self, el): count = next(self.counter) heapq.heappush(self._queue, (el, count))
push
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, el):\n count = next(self.counter)\n heapq.heappush(self._queue, (el, count))\n", "role": "user" }, { "content": "push", "role": "assistant" } ]
def x(self): sc = self._p.space_coverage gc = self._p.grain_coverage if sc and gc: if parse_to_gvid(gc[0]).level == 'state' and parse_to_gvid(sc[0] ).level == 'state': return parse_to_gvid(sc[0]).geo_name else: return '{} in {}'.format(parse_to_gvid(gc[0]).level_plural. title(), parse_to_gvid(sc[0]).geo_name) elif sc: return parse_to_gvid(sc[0]).geo_name.title() elif sc: return parse_to_gvid(gc[0]).level_plural.title() else: return ''
geo_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n sc = self._p.space_coverage\n gc = self._p.grain_coverage\n if sc and gc:\n if parse_to_gvid(gc[0]).level == 'state' and parse_to_gvid(sc[0]\n ).level == 'state':\n return parse_to_gvid(sc[0]).geo_name\n else:\n return '{} in {}'.format(parse_to_gvid(gc[0]).level_plural.\n title(), parse_to_gvid(sc[0]).geo_name)\n elif sc:\n return parse_to_gvid(sc[0]).geo_name.title()\n elif sc:\n return parse_to_gvid(gc[0]).level_plural.title()\n else:\n return ''\n", "role": "user" }, { "content": "geo_description", "role": "assistant" } ]
def x(self): tc = [t for t in self._p.time_coverage if t] if not tc: return '' mn = min(tc) mx = max(tc) if not mn and not mx: return '' elif mn == mx: return mn else: return '{} to {}'.format(mn, mx)
time_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n tc = [t for t in self._p.time_coverage if t]\n if not tc:\n return ''\n mn = min(tc)\n mx = max(tc)\n if not mn and not mx:\n return ''\n elif mn == mx:\n return mn\n else:\n return '{} to {}'.format(mn, mx)\n", "role": "user" }, { "content": "time_description", "role": "assistant" } ]
def x(self): gd = self.geo_description td = self.time_description if gd and td: return '{}, {}. {} Rows.'.format(gd, td, self._p.count) elif gd: return '{}. {} Rows.'.format(gd, self._p.count) elif td: return '{}. {} Rows.'.format(td, self._p.count) else: return '{} Rows.'.format(self._p.count)
sub_description
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n gd = self.geo_description\n td = self.time_description\n if gd and td:\n return '{}, {}. {} Rows.'.format(gd, td, self._p.count)\n elif gd:\n return '{}. {} Rows.'.format(gd, self._p.count)\n elif td:\n return '{}. {} Rows.'.format(td, self._p.count)\n else:\n return '{} Rows.'.format(self._p.count)\n", "role": "user" }, { "content": "sub_description", "role": "assistant" } ]
def x(self): if self.dataset is None: s = object_session(self) ds = s.query(Dataset).filter(Dataset.id_ == self.d_id).one() else: ds = self.dataset d = {'id': self.id, 'vid': self.vid, 'name': self.name, 'vname': self. vname, 'ref': self.ref, 'space': self.space, 'time': self.time, 'table': self.table_name, 'grain': self.grain, 'variant': self. variant, 'segment': self.segment, 'format': self.format if self. format else 'db'} return PartitionIdentity.from_dict(dict(list(ds.dict.items()) + list(d. items())))
identity
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if self.dataset is None:\n s = object_session(self)\n ds = s.query(Dataset).filter(Dataset.id_ == self.d_id).one()\n else:\n ds = self.dataset\n d = {'id': self.id, 'vid': self.vid, 'name': self.name, 'vname': self.\n vname, 'ref': self.ref, 'space': self.space, 'time': self.time,\n 'table': self.table_name, 'grain': self.grain, 'variant': self.\n variant, 'segment': self.segment, 'format': self.format if self.\n format else 'db'}\n return PartitionIdentity.from_dict(dict(list(ds.dict.items()) + list(d.\n items())))\n", "role": "user" }, { "content": "identity", "role": "assistant" } ]
def x(self, stats): from ambry.util.datestimes import expand_to_years scov = set() tcov = set() grains = set() def x(gvid): try: return parse_to_gvid(gvid).summarize() except: return None def x(values, column): parsed = [] for gvid in values: if gvid is None or gvid == 'None': continue try: parsed.append(parse_to_gvid(gvid)) except ValueError as e: if self._bundle: self._bundle.warn( 'While analyzing geo coverage in final partition stage, ' + "Failed to parse gvid '{}' in {}.{}: {}".format( str(gvid), column.table.name, column.name, e)) try: return isimplify(parsed) except: return None def x(year): try: return int(year) except: return None for c in self.table.columns: if c.name not in stats: continue try: if stats[c.name].is_gvid or stats[c.name].is_geoid: scov |= set(x for x in simplifiy_maybe(stats[c.name]. uniques, c)) grains |= set(summarize_maybe(gvid) for gvid in stats[c. name].uniques) elif stats[c.name].is_year: tcov |= set(int_maybe(x) for x in stats[c.name].uniques) elif stats[c.name].is_date: try: tcov |= set(parser.parse(x, fuzzy=True).year if isinstance(x, string_types) else x.year for x in stats[c.name].uniques) except ValueError: pass except Exception as e: self._bundle.error( "Failed to set coverage for column '{}', partition '{}': {}" .format(c.name, self.identity.vname, e)) raise if 'source_data' in self.data: for source_name, source in list(self.data['source_data'].items()): scov.add(self.parse_gvid_or_place(source['space'])) if self.identity.space: try: scov.add(self.parse_gvid_or_place(self.identity.space)) except ValueError: pass self.space_coverage = sorted([str(x) for x in scov if bool(x) and x.sl in (10, 40, 50, 60, 160, 400)]) if 'source_data' in self.data: for source_name, source in list(self.data['source_data'].items()): if 'time' in source: for year in expand_to_years(source['time']): if year: tcov.add(year) if self.identity.name.time: for year in expand_to_years(self.identity.name.time): if year: tcov.add(year) self.time_coverage = [t for t in tcov if t] if 'source_data' in self.data: for source_name, source in list(self.data['source_data'].items()): if 'grain' in source: grains.add(source['grain']) self.grain_coverage = sorted(str(g) for g in grains if g)
set_coverage
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, stats):\n from ambry.util.datestimes import expand_to_years\n scov = set()\n tcov = set()\n grains = set()\n\n def x(gvid):\n try:\n return parse_to_gvid(gvid).summarize()\n except:\n return None\n\n def x(values, column):\n parsed = []\n for gvid in values:\n if gvid is None or gvid == 'None':\n continue\n try:\n parsed.append(parse_to_gvid(gvid))\n except ValueError as e:\n if self._bundle:\n self._bundle.warn(\n 'While analyzing geo coverage in final partition stage, '\n + \"Failed to parse gvid '{}' in {}.{}: {}\".format(\n str(gvid), column.table.name, column.name, e))\n try:\n return isimplify(parsed)\n except:\n return None\n\n def x(year):\n try:\n return int(year)\n except:\n return None\n for c in self.table.columns:\n if c.name not in stats:\n continue\n try:\n if stats[c.name].is_gvid or stats[c.name].is_geoid:\n scov |= set(x for x in simplifiy_maybe(stats[c.name].\n uniques, c))\n grains |= set(summarize_maybe(gvid) for gvid in stats[c.\n name].uniques)\n elif stats[c.name].is_year:\n tcov |= set(int_maybe(x) for x in stats[c.name].uniques)\n elif stats[c.name].is_date:\n try:\n tcov |= set(parser.parse(x, fuzzy=True).year if\n isinstance(x, string_types) else x.year for x in\n stats[c.name].uniques)\n except ValueError:\n pass\n except Exception as e:\n self._bundle.error(\n \"Failed to set coverage for column '{}', partition '{}': {}\"\n .format(c.name, self.identity.vname, e))\n raise\n if 'source_data' in self.data:\n for source_name, source in list(self.data['source_data'].items()):\n scov.add(self.parse_gvid_or_place(source['space']))\n if self.identity.space:\n try:\n scov.add(self.parse_gvid_or_place(self.identity.space))\n except ValueError:\n pass\n self.space_coverage = sorted([str(x) for x in scov if bool(x) and x.sl in\n (10, 40, 50, 60, 160, 400)])\n if 'source_data' in self.data:\n for source_name, source in list(self.data['source_data'].items()):\n if 'time' in source:\n for year in expand_to_years(source['time']):\n if year:\n tcov.add(year)\n if self.identity.name.time:\n for year in expand_to_years(self.identity.name.time):\n if year:\n tcov.add(year)\n self.time_coverage = [t for t in tcov if t]\n if 'source_data' in self.data:\n for source_name, source in list(self.data['source_data'].items()):\n if 'grain' in source:\n grains.add(source['grain'])\n self.grain_coverage = sorted(str(g) for g in grains if g)\n", "role": "user" }, { "content": "set_coverage", "role": "assistant" } ]
def x(self): d = self.dict def x(c): d = c.dict d['stats'] = [s.dict for s in c.stats] return d d['table'] = self.table.dict d['table']['columns'] = [aug_col(c) for c in self.table.columns] return d
detail_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n d = self.dict\n\n def x(c):\n d = c.dict\n d['stats'] = [s.dict for s in c.stats]\n return d\n d['table'] = self.table.dict\n d['table']['columns'] = [aug_col(c) for c in self.table.columns]\n return d\n", "role": "user" }, { "content": "detail_dict", "role": "assistant" } ]
def x(self): from ambry_sources import MPRowsFile from fs.errors import ResourceNotFoundError from ambry.orm.exc import NotFoundError try: return MPRowsFile(self._bundle.build_fs, self.cache_key) except ResourceNotFoundError: raise NotFoundError( 'Could not locate data file for partition {} (local)'.format( self.identity.fqname))
local_datafile
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry_sources import MPRowsFile\n from fs.errors import ResourceNotFoundError\n from ambry.orm.exc import NotFoundError\n try:\n return MPRowsFile(self._bundle.build_fs, self.cache_key)\n except ResourceNotFoundError:\n raise NotFoundError(\n 'Could not locate data file for partition {} (local)'.format(\n self.identity.fqname))\n", "role": "user" }, { "content": "local_datafile", "role": "assistant" } ]
def x(self): from ambry.exc import NotFoundError ds = self.dataset if 'remote_name' not in ds.data: raise NotFoundError('Could not determine remote for partition: {}'. format(self.identity.fqname)) return self._bundle.library.remote(ds.data['remote_name'])
remote
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.exc import NotFoundError\n ds = self.dataset\n if 'remote_name' not in ds.data:\n raise NotFoundError('Could not determine remote for partition: {}'.\n format(self.identity.fqname))\n return self._bundle.library.remote(ds.data['remote_name'])\n", "role": "user" }, { "content": "remote", "role": "assistant" } ]
def x(self): from ambry.orm.exc import NotFoundError try: if self.local_datafile.exists: return True except NotFoundError: pass return False
is_local
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.orm.exc import NotFoundError\n try:\n if self.local_datafile.exists:\n return True\n except NotFoundError:\n pass\n return False\n", "role": "user" }, { "content": "is_local", "role": "assistant" } ]
def x(self, ps=None): from filelock import FileLock from ambry.util import ensure_dir_exists from ambry_sources import MPRowsFile from fs.errors import ResourceNotFoundError if self.is_local: return local = self._bundle.build_fs b = self._bundle.library.bundle(self.identity.as_dataset().vid) remote = self._bundle.library.remote(b) lock_path = local.getsyspath(self.cache_key + '.lock') ensure_dir_exists(lock_path) lock = FileLock(lock_path) if ps: ps.add_update(message='Localizing {}'.format(self.identity.name), partition=self, item_type='bytes', state='downloading') if ps: def x(bts): if ps.rec.item_total is None: ps.rec.item_count = 0 if not ps.rec.data: ps.rec.data = {} return self item_count = ps.rec.item_count + bts ps.rec.data['updates'] = ps.rec.data.get('updates', 0) + 1 if ps.rec.data['updates'] % 32 == 1: ps.update(message='Localizing {}'.format(self.identity.name ), item_count=item_count) else: from ambry.bundle.process import call_interval @call_interval(5) def x(bts): self._bundle.log('Localizing {}. {} bytes downloaded'.format( self.vname, bts)) def x(e): raise e with lock: if self.is_local: return self try: with remote.fs.open(self.cache_key + MPRowsFile.EXTENSION, 'rb' ) as f: event = local.setcontents_async(self.cache_key + MPRowsFile .EXTENSION, f, progress_callback=progress, error_callback=exception_cb) event.wait() if ps: ps.update_done() except ResourceNotFoundError as e: from ambry.orm.exc import NotFoundError raise NotFoundError("Failed to get MPRfile '{}' from {}: {} ". format(self.cache_key, remote.fs, e)) return self
localize
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, ps=None):\n from filelock import FileLock\n from ambry.util import ensure_dir_exists\n from ambry_sources import MPRowsFile\n from fs.errors import ResourceNotFoundError\n if self.is_local:\n return\n local = self._bundle.build_fs\n b = self._bundle.library.bundle(self.identity.as_dataset().vid)\n remote = self._bundle.library.remote(b)\n lock_path = local.getsyspath(self.cache_key + '.lock')\n ensure_dir_exists(lock_path)\n lock = FileLock(lock_path)\n if ps:\n ps.add_update(message='Localizing {}'.format(self.identity.name),\n partition=self, item_type='bytes', state='downloading')\n if ps:\n\n def x(bts):\n if ps.rec.item_total is None:\n ps.rec.item_count = 0\n if not ps.rec.data:\n ps.rec.data = {}\n return self\n item_count = ps.rec.item_count + bts\n ps.rec.data['updates'] = ps.rec.data.get('updates', 0) + 1\n if ps.rec.data['updates'] % 32 == 1:\n ps.update(message='Localizing {}'.format(self.identity.name\n ), item_count=item_count)\n else:\n from ambry.bundle.process import call_interval\n\n @call_interval(5)\n def x(bts):\n self._bundle.log('Localizing {}. {} bytes downloaded'.format(\n self.vname, bts))\n\n def x(e):\n raise e\n with lock:\n if self.is_local:\n return self\n try:\n with remote.fs.open(self.cache_key + MPRowsFile.EXTENSION, 'rb'\n ) as f:\n event = local.setcontents_async(self.cache_key + MPRowsFile\n .EXTENSION, f, progress_callback=progress,\n error_callback=exception_cb)\n event.wait()\n if ps:\n ps.update_done()\n except ResourceNotFoundError as e:\n from ambry.orm.exc import NotFoundError\n raise NotFoundError(\"Failed to get MPRfile '{}' from {}: {} \".\n format(self.cache_key, remote.fs, e))\n return self\n", "role": "user" }, { "content": "localize", "role": "assistant" } ]
def x(self): from ambry.orm.exc import NotFoundError from fs.errors import ResourceNotFoundError """The reader for the datafile""" try: return self.datafile.reader except ResourceNotFoundError: raise NotFoundError("Failed to find partition file, '{}' ".format( self.datafile.path))
reader
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.orm.exc import NotFoundError\n from fs.errors import ResourceNotFoundError\n \"\"\"The reader for the datafile\"\"\"\n try:\n return self.datafile.reader\n except ResourceNotFoundError:\n raise NotFoundError(\"Failed to find partition file, '{}' \".format(\n self.datafile.path))\n", "role": "user" }, { "content": "reader", "role": "assistant" } ]
def x(self, predicate=None, headers=None): with self.reader as r: for row in r.select(predicate, headers): yield row
select
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, predicate=None, headers=None):\n with self.reader as r:\n for row in r.select(predicate, headers):\n yield row\n", "role": "user" }, { "content": "select", "role": "assistant" } ]
def x(self): if isinstance(self, PartitionProxy): return AnalysisPartition(self._obj) else: return AnalysisPartition(self)
analysis
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if isinstance(self, PartitionProxy):\n return AnalysisPartition(self._obj)\n else:\n return AnalysisPartition(self)\n", "role": "user" }, { "content": "analysis", "role": "assistant" } ]
def x(self): if isinstance(self, PartitionProxy): return MeasureDimensionPartition(self._obj) else: return MeasureDimensionPartition(self)
measuredim
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n if isinstance(self, PartitionProxy):\n return MeasureDimensionPartition(self._obj)\n else:\n return MeasureDimensionPartition(self)\n", "role": "user" }, { "content": "measuredim", "role": "assistant" } ]
def x(self, sequence_id=None): if sequence_id: self.sequence_id = sequence_id self._set_ids(force=True) if self.dataset: self._update_names()
update_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, sequence_id=None):\n if sequence_id:\n self.sequence_id = sequence_id\n self._set_ids(force=True)\n if self.dataset:\n self._update_names()\n", "role": "user" }, { "content": "update_id", "role": "assistant" } ]
def x(self): d = dict(table=self.table_name, time=self.time, space=self.space, grain =self.grain, variant=self.variant, segment=self.segment) assert self.dataset name = PartialPartitionName(**d).promote(self.dataset.identity.name) self.name = str(name.name) self.vname = str(name.vname) self.cache_key = name.cache_key self.fqname = str(self.identity.fqname)
update_names
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n d = dict(table=self.table_name, time=self.time, space=self.space, grain\n =self.grain, variant=self.variant, segment=self.segment)\n assert self.dataset\n name = PartialPartitionName(**d).promote(self.dataset.identity.name)\n self.name = str(name.name)\n self.vname = str(name.vname)\n self.cache_key = name.cache_key\n self.fqname = str(self.identity.fqname)\n", "role": "user" }, { "content": "update_names", "role": "assistant" } ]
def x(mapper, conn, target): target._set_ids() if (target.name and target.vname and target.cache_key and target.fqname and not target.dataset): return Partition.before_update(mapper, conn, target)
before_insert
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(mapper, conn, target):\n target._set_ids()\n if (target.name and target.vname and target.cache_key and target.fqname and\n not target.dataset):\n return\n Partition.before_update(mapper, conn, target)\n", "role": "user" }, { "content": "before_insert", "role": "assistant" } ]
def x(self, predicate=None, filtered_columns=None, columns=None, df_class=None ): from operator import itemgetter from ambry.pands import AmbryDataFrame df_class = df_class or AmbryDataFrame if columns: ig = itemgetter(*columns) else: ig = None columns = self.table.header if filtered_columns: def x(v): from six import string_types if isinstance(v, string_types): return '"{}"'.format(v) else: return v code = ' and '.join('row.{} == {}'.format(k, maybe_quote(v)) for k, v in filtered_columns.items()) predicate = eval('lambda row: {}'.format(code)) if predicate: def x(): for row in self.reader: if predicate(row): if ig: yield ig(row) else: yield row.dict df = df_class(yielder(), columns=columns, partition=self.measuredim) return df else: def x(): for row in self.reader: yield row.values() columns = [c for c in self.table.header if c in columns] return df_class(yielder(), columns=columns, partition=self.measuredim)
dataframe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, predicate=None, filtered_columns=None, columns=None, df_class=None\n ):\n from operator import itemgetter\n from ambry.pands import AmbryDataFrame\n df_class = df_class or AmbryDataFrame\n if columns:\n ig = itemgetter(*columns)\n else:\n ig = None\n columns = self.table.header\n if filtered_columns:\n\n def x(v):\n from six import string_types\n if isinstance(v, string_types):\n return '\"{}\"'.format(v)\n else:\n return v\n code = ' and '.join('row.{} == {}'.format(k, maybe_quote(v)) for k,\n v in filtered_columns.items())\n predicate = eval('lambda row: {}'.format(code))\n if predicate:\n\n def x():\n for row in self.reader:\n if predicate(row):\n if ig:\n yield ig(row)\n else:\n yield row.dict\n df = df_class(yielder(), columns=columns, partition=self.measuredim)\n return df\n else:\n\n def x():\n for row in self.reader:\n yield row.values()\n columns = [c for c in self.table.header if c in columns]\n return df_class(yielder(), columns=columns, partition=self.measuredim)\n", "role": "user" }, { "content": "dataframe", "role": "assistant" } ]
def x(self, simplify=None, predicate=None): from shapely.wkt import loads if not predicate: predicate = lambda row: True if simplify: return [loads(row.geometry).simplify(simplify) for row in self if predicate(row)] else: return [loads(row.geometry) for row in self if predicate(row)]
shapes
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, simplify=None, predicate=None):\n from shapely.wkt import loads\n if not predicate:\n predicate = lambda row: True\n if simplify:\n return [loads(row.geometry).simplify(simplify) for row in self if\n predicate(row)]\n else:\n return [loads(row.geometry) for row in self if predicate(row)]\n", "role": "user" }, { "content": "shapes", "role": "assistant" } ]
def x(self, basemap, simplify=None, predicate=None, args_f=None, **kwargs): from descartes import PolygonPatch from shapely.wkt import loads from shapely.ops import transform if not predicate: predicate = lambda row: True def x(x, y, z=None): return basemap(x, y) def x(shape, row): args = dict(kwargs.items()) if args_f: args.update(args_f(row)) return PolygonPatch(transform(map_xform, shape), **args) def x(row): if simplify: shape = loads(row.geometry).simplify(simplify) else: shape = loads(row.geometry) if shape.geom_type == 'MultiPolygon': for subshape in shape.geoms: yield make_patch(subshape, row) else: yield make_patch(shape, row) return [patch for row in self if predicate(row) for patch in yield_patches(row)]
patches
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, basemap, simplify=None, predicate=None, args_f=None, **kwargs):\n from descartes import PolygonPatch\n from shapely.wkt import loads\n from shapely.ops import transform\n if not predicate:\n predicate = lambda row: True\n\n def x(x, y, z=None):\n return basemap(x, y)\n\n def x(shape, row):\n args = dict(kwargs.items())\n if args_f:\n args.update(args_f(row))\n return PolygonPatch(transform(map_xform, shape), **args)\n\n def x(row):\n if simplify:\n shape = loads(row.geometry).simplify(simplify)\n else:\n shape = loads(row.geometry)\n if shape.geom_type == 'MultiPolygon':\n for subshape in shape.geoms:\n yield make_patch(subshape, row)\n else:\n yield make_patch(shape, row)\n return [patch for row in self if predicate(row) for patch in\n yield_patches(row)]\n", "role": "user" }, { "content": "patches", "role": "assistant" } ]
def x(self): from ambry.valuetype.core import ROLE return [c for c in self.columns if c.role == ROLE.MEASURE]
measures
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from ambry.valuetype.core import ROLE\n return [c for c in self.columns if c.role == ROLE.MEASURE]\n", "role": "user" }, { "content": "measures", "role": "assistant" } ]
def x(self, vid): from ambry.orm import Column if isinstance(vid, PartitionColumn): return vid elif isinstance(vid, Column): return PartitionColumn(vid) else: return PartitionColumn(self.table.column(vid), self)
measure
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, vid):\n from ambry.orm import Column\n if isinstance(vid, PartitionColumn):\n return vid\n elif isinstance(vid, Column):\n return PartitionColumn(vid)\n else:\n return PartitionColumn(self.table.column(vid), self)\n", "role": "user" }, { "content": "measure", "role": "assistant" } ]
def x(self, measure, p_dim, s_dim=None, filters={}, df_class=None): import numpy as np measure = self.measure(measure) p_dim = self.dimension(p_dim) assert p_dim if s_dim: s_dim = self.dimension(s_dim) columns = set([measure.name, p_dim.name]) if p_dim.label: if p_dim.geoid: columns.add(p_dim.geoid.name) columns.add(p_dim.label.name) if s_dim: columns.add(s_dim.name) if s_dim.label: columns.add(s_dim.label.name) def x(v): from six import string_types if isinstance(v, string_types): return '"{}"'.format(v) else: return v if filters: selected_filters = [] for k, v in filters.items(): if isinstance(v, dict): v = v.keys()[0] selected_filters.append('row.{} == {}'.format(k, maybe_quote(v))) code = ' and '.join(selected_filters) predicate = eval('lambda row: {}'.format(code)) else: code = None def x(row): return True df = self.analysis.dataframe(predicate, columns=columns, df_class=df_class) if df is None or df.empty or len(df) == 0: return None df['_count'] = 1 def x(x): return ', '.join(set(str(e) for e in x)) agg = {'_count': 'count'} for col_name in columns: c = self.column(col_name) if c.name == p_dim.name or s_dim and c.name == s_dim.name: continue if c.parent and (c.parent == p_dim.name or s_dim and c.parent == s_dim.name): continue if c.is_measure: agg[c.name] = np.mean if c.is_dimension: agg[c.name] = aggregate_string plot_axes = [p_dim.name] if s_dim: plot_axes.append(s_dim.name) df = df.groupby(list(columns - set([measure.name]))).agg(agg).reset_index() df._metadata = ['plot_axes', 'filtered', 'floating', 'labels', 'dimension_set', 'measure'] df.plot_axes = [c for c in plot_axes] df.filtered = filters df.floating = list(set(c.name for c in self.primary_dimensions) - set( df.filtered.keys()) - set(df.plot_axes)) df.labels = [(self.column(c).label.name if self.column(c).label else c) for c in df.plot_axes] df.dimension_set = self.dimension_set(p_dim, s_dim=s_dim) df.measure = measure.name def x(self): yield ['id'] + list(df.columns) for t in df.itertuples(): yield list(t) df.__class__.rows = property(rows) return df
dataframe
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, measure, p_dim, s_dim=None, filters={}, df_class=None):\n import numpy as np\n measure = self.measure(measure)\n p_dim = self.dimension(p_dim)\n assert p_dim\n if s_dim:\n s_dim = self.dimension(s_dim)\n columns = set([measure.name, p_dim.name])\n if p_dim.label:\n if p_dim.geoid:\n columns.add(p_dim.geoid.name)\n columns.add(p_dim.label.name)\n if s_dim:\n columns.add(s_dim.name)\n if s_dim.label:\n columns.add(s_dim.label.name)\n\n def x(v):\n from six import string_types\n if isinstance(v, string_types):\n return '\"{}\"'.format(v)\n else:\n return v\n if filters:\n selected_filters = []\n for k, v in filters.items():\n if isinstance(v, dict):\n v = v.keys()[0]\n selected_filters.append('row.{} == {}'.format(k, maybe_quote(v)))\n code = ' and '.join(selected_filters)\n predicate = eval('lambda row: {}'.format(code))\n else:\n code = None\n\n def x(row):\n return True\n df = self.analysis.dataframe(predicate, columns=columns, df_class=df_class)\n if df is None or df.empty or len(df) == 0:\n return None\n df['_count'] = 1\n\n def x(x):\n return ', '.join(set(str(e) for e in x))\n agg = {'_count': 'count'}\n for col_name in columns:\n c = self.column(col_name)\n if c.name == p_dim.name or s_dim and c.name == s_dim.name:\n continue\n if c.parent and (c.parent == p_dim.name or s_dim and c.parent ==\n s_dim.name):\n continue\n if c.is_measure:\n agg[c.name] = np.mean\n if c.is_dimension:\n agg[c.name] = aggregate_string\n plot_axes = [p_dim.name]\n if s_dim:\n plot_axes.append(s_dim.name)\n df = df.groupby(list(columns - set([measure.name]))).agg(agg).reset_index()\n df._metadata = ['plot_axes', 'filtered', 'floating', 'labels',\n 'dimension_set', 'measure']\n df.plot_axes = [c for c in plot_axes]\n df.filtered = filters\n df.floating = list(set(c.name for c in self.primary_dimensions) - set(\n df.filtered.keys()) - set(df.plot_axes))\n df.labels = [(self.column(c).label.name if self.column(c).label else c) for\n c in df.plot_axes]\n df.dimension_set = self.dimension_set(p_dim, s_dim=s_dim)\n df.measure = measure.name\n\n def x(self):\n yield ['id'] + list(df.columns)\n for t in df.itertuples():\n yield list(t)\n df.__class__.rows = property(rows)\n return df\n", "role": "user" }, { "content": "dataframe", "role": "assistant" } ]
def x(self, p_dim, s_dim=None, dimensions=None, extant=set()): if not dimensions: dimensions = self.primary_dimensions key = p_dim.name if s_dim: key += '/' + s_dim.name if key in extant or p_dim == s_dim: return if s_dim and s_dim.valuetype_class.is_geo(): return extant.add(key) filtered = {} for d in dimensions: if d != p_dim and d != s_dim: filtered[d.name] = d.pstats.uvalues.keys() if p_dim.valuetype_class.is_time(): value_type = 'time' chart_type = 'line' elif p_dim.valuetype_class.is_geo(): value_type = 'geo' chart_type = 'map' else: value_type = 'general' chart_type = 'bar' return dict(key=key, p_dim=p_dim.name, p_dim_type=value_type, p_label= p_dim.label_or_self.name, s_dim=s_dim.name if s_dim else None, s_label=s_dim.label_or_self.name if s_dim else None, filters= filtered, chart_type=chart_type)
dimension_set
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, p_dim, s_dim=None, dimensions=None, extant=set()):\n if not dimensions:\n dimensions = self.primary_dimensions\n key = p_dim.name\n if s_dim:\n key += '/' + s_dim.name\n if key in extant or p_dim == s_dim:\n return\n if s_dim and s_dim.valuetype_class.is_geo():\n return\n extant.add(key)\n filtered = {}\n for d in dimensions:\n if d != p_dim and d != s_dim:\n filtered[d.name] = d.pstats.uvalues.keys()\n if p_dim.valuetype_class.is_time():\n value_type = 'time'\n chart_type = 'line'\n elif p_dim.valuetype_class.is_geo():\n value_type = 'geo'\n chart_type = 'map'\n else:\n value_type = 'general'\n chart_type = 'bar'\n return dict(key=key, p_dim=p_dim.name, p_dim_type=value_type, p_label=\n p_dim.label_or_self.name, s_dim=s_dim.name if s_dim else None,\n s_label=s_dim.label_or_self.name if s_dim else None, filters=\n filtered, chart_type=chart_type)\n", "role": "user" }, { "content": "dimension_set", "role": "assistant" } ]
def x(self): for c in self.table.columns: if c.parent == self.name and 'label' in c.valuetype: return PartitionColumn(c, self._partition)
label
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n for c in self.table.columns:\n if c.parent == self.name and 'label' in c.valuetype:\n return PartitionColumn(c, self._partition)\n", "role": "user" }, { "content": "label", "role": "assistant" } ]
def x(self): from operator import itemgetter card = self.pstats.nuniques if self.label: ig = itemgetter(self.name, self.label.name) elif self.pstats.nuniques < MAX_LABELS: ig = itemgetter(self.name, self.name) else: return {} label_set = set() for row in self._partition: label_set.add(ig(row)) if len(label_set) >= card: break d = dict(label_set) assert len(d) == len(label_set) return d
value_labels
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n from operator import itemgetter\n card = self.pstats.nuniques\n if self.label:\n ig = itemgetter(self.name, self.label.name)\n elif self.pstats.nuniques < MAX_LABELS:\n ig = itemgetter(self.name, self.name)\n else:\n return {}\n label_set = set()\n for row in self._partition:\n label_set.add(ig(row))\n if len(label_set) >= card:\n break\n d = dict(label_set)\n assert len(d) == len(label_set)\n return d\n", "role": "user" }, { "content": "value_labels", "role": "assistant" } ]
def x(page_size=200, page_index=0, sort='', q=''): data = list_build_configuration_sets_raw(page_size, page_index, sort, q) if data: return utils.format_json_list(data)
list_build_configuration_sets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(page_size=200, page_index=0, sort='', q=''):\n data = list_build_configuration_sets_raw(page_size, page_index, sort, q)\n if data:\n return utils.format_json_list(data)\n", "role": "user" }, { "content": "list_build_configuration_sets", "role": "assistant" } ]
def x(**kwargs): config_set = _create_build_config_set_object(**kwargs) response = utils.checked_api_call(pnc_api.build_group_configs, 'create_new', body=config_set) if response: return response.content
create_build_configuration_set_raw
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(**kwargs):\n config_set = _create_build_config_set_object(**kwargs)\n response = utils.checked_api_call(pnc_api.build_group_configs,\n 'create_new', body=config_set)\n if response:\n return response.content\n", "role": "user" }, { "content": "create_build_configuration_set_raw", "role": "assistant" } ]