source_code
stringlengths
45
7.24k
function_name
stringlengths
1
111
messages
list
def x(d: dict) ->dict: d_out = {} for k, v in d.items(): if v in d_out: d_out[v].append(k) else: d_out[v] = [k] return d_out
group_dict_by_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(d: dict) ->dict:\n d_out = {}\n for k, v in d.items():\n if v in d_out:\n d_out[v].append(k)\n else:\n d_out[v] = [k]\n return d_out\n", "role": "user" }, { "content": "group_dict_by_value", "role": "assistant" } ]
def x(code: str, exclude_variable: Union[set, None]=None, jsonable_parameter: bool=True) ->tuple: if exclude_variable is None: exclude_variable = set() else: exclude_variable = copy.deepcopy(exclude_variable) root = ast.parse(code) store_variable_name = set() assign_only = True dict_parameter = {} for node in ast.iter_child_nodes(root): if isinstance(node, ast.Assign): for assign_node in ast.walk(node): if isinstance(assign_node, ast.Name): if isinstance(assign_node.ctx, ast.Store): if jsonable_parameter is False: store_variable_name |= {assign_node.id} else: exclude_variable |= {assign_node.id} _is_literal_eval, _value = is_literal_eval(node.value) if jsonable_parameter is True: for assign_node in ast.iter_child_nodes(node): if isinstance(assign_node, ast.Tuple): i = 0 for assign_tuple_node in ast.iter_child_nodes( assign_node): if isinstance(assign_tuple_node, ast.Name): if isinstance(_value, collections.Iterable ) and is_jsonable(_value[i] ) and _is_literal_eval: dict_parameter[assign_tuple_node.id ] = _value[i] store_variable_name |= {assign_tuple_node .id} else: exclude_variable |= {assign_tuple_node.id} i += 1 elif isinstance(assign_node, ast.Name): if is_jsonable(_value) and _is_literal_eval: dict_parameter[assign_node.id] = _value store_variable_name |= {assign_node.id} else: exclude_variable |= {assign_node.id} elif isinstance(node, ast.AugAssign): for assign_node in ast.walk(node): if isinstance(assign_node, ast.Name): exclude_variable |= {assign_node.id} elif isinstance(node, (ast.FunctionDef, ast.ClassDef)): assign_only = False exclude_variable |= {node.name} elif isinstance(node, ast.Import): assign_only = False for node1 in ast.iter_child_nodes(node): if node1.asname is not None: exclude_variable |= {node1.asname} else: exclude_variable |= {node1.name} elif isinstance(node, ast.ImportFrom): assign_only = False for node1 in ast.iter_child_nodes(node): if node1.asname is not None: exclude_variable |= {node1.asname} else: exclude_variable |= {node1.name} else: assign_only = False if assign_only is True: possible_parameter = store_variable_name - exclude_variable if jsonable_parameter is True: dict_parameter = {k: dict_parameter[k] for k in possible_parameter} return (possible_parameter, store_variable_name | exclude_variable, dict_parameter) return set(), store_variable_name | exclude_variable, {}
variable_status
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(code: str, exclude_variable: Union[set, None]=None,\n jsonable_parameter: bool=True) ->tuple:\n if exclude_variable is None:\n exclude_variable = set()\n else:\n exclude_variable = copy.deepcopy(exclude_variable)\n root = ast.parse(code)\n store_variable_name = set()\n assign_only = True\n dict_parameter = {}\n for node in ast.iter_child_nodes(root):\n if isinstance(node, ast.Assign):\n for assign_node in ast.walk(node):\n if isinstance(assign_node, ast.Name):\n if isinstance(assign_node.ctx, ast.Store):\n if jsonable_parameter is False:\n store_variable_name |= {assign_node.id}\n else:\n exclude_variable |= {assign_node.id}\n _is_literal_eval, _value = is_literal_eval(node.value)\n if jsonable_parameter is True:\n for assign_node in ast.iter_child_nodes(node):\n if isinstance(assign_node, ast.Tuple):\n i = 0\n for assign_tuple_node in ast.iter_child_nodes(\n assign_node):\n if isinstance(assign_tuple_node, ast.Name):\n if isinstance(_value, collections.Iterable\n ) and is_jsonable(_value[i]\n ) and _is_literal_eval:\n dict_parameter[assign_tuple_node.id\n ] = _value[i]\n store_variable_name |= {assign_tuple_node\n .id}\n else:\n exclude_variable |= {assign_tuple_node.id}\n i += 1\n elif isinstance(assign_node, ast.Name):\n if is_jsonable(_value) and _is_literal_eval:\n dict_parameter[assign_node.id] = _value\n store_variable_name |= {assign_node.id}\n else:\n exclude_variable |= {assign_node.id}\n elif isinstance(node, ast.AugAssign):\n for assign_node in ast.walk(node):\n if isinstance(assign_node, ast.Name):\n exclude_variable |= {assign_node.id}\n elif isinstance(node, (ast.FunctionDef, ast.ClassDef)):\n assign_only = False\n exclude_variable |= {node.name}\n elif isinstance(node, ast.Import):\n assign_only = False\n for node1 in ast.iter_child_nodes(node):\n if node1.asname is not None:\n exclude_variable |= {node1.asname}\n else:\n exclude_variable |= {node1.name}\n elif isinstance(node, ast.ImportFrom):\n assign_only = False\n for node1 in ast.iter_child_nodes(node):\n if node1.asname is not None:\n exclude_variable |= {node1.asname}\n else:\n exclude_variable |= {node1.name}\n else:\n assign_only = False\n if assign_only is True:\n possible_parameter = store_variable_name - exclude_variable\n if jsonable_parameter is True:\n dict_parameter = {k: dict_parameter[k] for k in possible_parameter}\n return (possible_parameter, store_variable_name | exclude_variable,\n dict_parameter)\n return set(), store_variable_name | exclude_variable, {}\n", "role": "user" }, { "content": "variable_status", "role": "assistant" } ]
def x(name: str, start_marker: str=' (', end_marker: str=')') ->str: if start_marker == '': raise ValueError('start_marker can not be the empty string.') a = name start = len(a) - a[::-1].find(start_marker[::-1]) if a[len(a) - len(end_marker):len(a)] == end_marker and start < len(a ) - len(end_marker) and a[start - len(start_marker):start ] == start_marker and a[start:len(a) - len(end_marker)].isdigit(): old_int = int(a[start:len(a) - len(end_marker)]) new_int = old_int + 1 new_name = a[:start] + str(new_int) + end_marker else: new_name = a + start_marker + '1' + end_marker return new_name
increment_name
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name: str, start_marker: str=' (', end_marker: str=')') ->str:\n if start_marker == '':\n raise ValueError('start_marker can not be the empty string.')\n a = name\n start = len(a) - a[::-1].find(start_marker[::-1])\n if a[len(a) - len(end_marker):len(a)] == end_marker and start < len(a\n ) - len(end_marker) and a[start - len(start_marker):start\n ] == start_marker and a[start:len(a) - len(end_marker)].isdigit():\n old_int = int(a[start:len(a) - len(end_marker)])\n new_int = old_int + 1\n new_name = a[:start] + str(new_int) + end_marker\n else:\n new_name = a + start_marker + '1' + end_marker\n return new_name\n", "role": "user" }, { "content": "increment_name", "role": "assistant" } ]
def x(identifier, configuration=None): resourceview = ResourceView(configuration=configuration) result = resourceview._load_from_hdx('resource view', identifier) if result: return resourceview return None
read_from_hdx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(identifier, configuration=None):\n resourceview = ResourceView(configuration=configuration)\n result = resourceview._load_from_hdx('resource view', identifier)\n if result:\n return resourceview\n return None\n", "role": "user" }, { "content": "read_from_hdx", "role": "assistant" } ]
def x(identifier, configuration=None): resourceview = ResourceView(configuration=configuration) success, result = resourceview._read_from_hdx('resource view', identifier, 'id', ResourceView.actions()['list']) resourceviews = list() if success: for resourceviewdict in result: resourceview = ResourceView(resourceviewdict, configuration= configuration) resourceviews.append(resourceview) return resourceviews
get_all_for_resource
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(identifier, configuration=None):\n resourceview = ResourceView(configuration=configuration)\n success, result = resourceview._read_from_hdx('resource view',\n identifier, 'id', ResourceView.actions()['list'])\n resourceviews = list()\n if success:\n for resourceviewdict in result:\n resourceview = ResourceView(resourceviewdict, configuration=\n configuration)\n resourceviews.append(resourceview)\n return resourceviews\n", "role": "user" }, { "content": "get_all_for_resource", "role": "assistant" } ]
def x(self, log=False): update = False if 'id' in self.data and self._load_from_hdx('resource view', self.data ['id']): update = True elif 'resource_id' in self.data: resource_views = self.get_all_for_resource(self.data['resource_id']) for resource_view in resource_views: if self.data['title'] == resource_view['title']: self.old_data = self.data self.data = resource_view.data update = True break if update: if log: logger.warning('resource view exists. Updating %s' % self.data[ 'id']) self._merge_hdx_update('resource view', 'id') return update
update_resource_view
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, log=False):\n update = False\n if 'id' in self.data and self._load_from_hdx('resource view', self.data\n ['id']):\n update = True\n elif 'resource_id' in self.data:\n resource_views = self.get_all_for_resource(self.data['resource_id'])\n for resource_view in resource_views:\n if self.data['title'] == resource_view['title']:\n self.old_data = self.data\n self.data = resource_view.data\n update = True\n break\n if update:\n if log:\n logger.warning('resource view exists. Updating %s' % self.data[\n 'id'])\n self._merge_hdx_update('resource view', 'id')\n return update\n", "role": "user" }, { "content": "update_resource_view", "role": "assistant" } ]
def x(self): self.check_required_fields() if not self._update_resource_view(log=True): self._save_to_hdx('create', 'title')
create_in_hdx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n self.check_required_fields()\n if not self._update_resource_view(log=True):\n self._save_to_hdx('create', 'title')\n", "role": "user" }, { "content": "create_in_hdx", "role": "assistant" } ]
def x(self, resource_view): if isinstance(resource_view, str): if is_valid_uuid(resource_view) is False: raise HDXError('%s is not a valid resource view id!' % resource_view) resource_view = ResourceView.read_from_hdx(resource_view) if not isinstance(resource_view, dict) and not isinstance(resource_view, ResourceView): raise HDXError('%s is not a valid resource view!' % resource_view) for key in resource_view: if key not in ('id', 'resource_id', 'package_id'): self.data[key] = resource_view[key]
copy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, resource_view):\n if isinstance(resource_view, str):\n if is_valid_uuid(resource_view) is False:\n raise HDXError('%s is not a valid resource view id!' %\n resource_view)\n resource_view = ResourceView.read_from_hdx(resource_view)\n if not isinstance(resource_view, dict) and not isinstance(resource_view,\n ResourceView):\n raise HDXError('%s is not a valid resource view!' % resource_view)\n for key in resource_view:\n if key not in ('id', 'resource_id', 'package_id'):\n self.data[key] = resource_view[key]\n", "role": "user" }, { "content": "copy", "role": "assistant" } ]
def x(configuration=None, url=None, keycolumn=5, failchained=True): if not Tags._tags_dict: if configuration is None: configuration = Configuration.read() with Download(full_agent=configuration.get_user_agent()) as downloader: if url is None: url = configuration['tags_cleanup_url'] Tags._tags_dict = downloader.download_tabular_rows_as_dicts(url, keycolumn=keycolumn) keys = Tags._tags_dict.keys() chainerror = False for i, tag in enumerate(keys): whattodo = Tags._tags_dict[tag] action = whattodo[u'action'] final_tags = whattodo[u'final tags (semicolon separated)'] for final_tag in final_tags.split(';'): if final_tag in keys: index = list(keys).index(final_tag) if index != i: whattodo2 = Tags._tags_dict[final_tag] action2 = whattodo2[u'action'] if action2 != 'OK' and action2 != 'Other': final_tags2 = whattodo2[ u'final tags (semicolon separated)'] if final_tag not in final_tags2.split(';'): chainerror = True if failchained: logger.error( 'Chained rules: %s (%s -> %s) | %s (%s -> %s)' % (action, tag, final_tags, action2, final_tag, final_tags2)) if failchained and chainerror: raise ChainRuleError('Chained rules for tags detected!') Tags._wildcard_tags = list() for tag in Tags._tags_dict: if '*' in tag: Tags._wildcard_tags.append(tag) return Tags._tags_dict, Tags._wildcard_tags
tagscleanupdicts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(configuration=None, url=None, keycolumn=5, failchained=True):\n if not Tags._tags_dict:\n if configuration is None:\n configuration = Configuration.read()\n with Download(full_agent=configuration.get_user_agent()) as downloader:\n if url is None:\n url = configuration['tags_cleanup_url']\n Tags._tags_dict = downloader.download_tabular_rows_as_dicts(url,\n keycolumn=keycolumn)\n keys = Tags._tags_dict.keys()\n chainerror = False\n for i, tag in enumerate(keys):\n whattodo = Tags._tags_dict[tag]\n action = whattodo[u'action']\n final_tags = whattodo[u'final tags (semicolon separated)']\n for final_tag in final_tags.split(';'):\n if final_tag in keys:\n index = list(keys).index(final_tag)\n if index != i:\n whattodo2 = Tags._tags_dict[final_tag]\n action2 = whattodo2[u'action']\n if action2 != 'OK' and action2 != 'Other':\n final_tags2 = whattodo2[\n u'final tags (semicolon separated)']\n if final_tag not in final_tags2.split(';'):\n chainerror = True\n if failchained:\n logger.error(\n 'Chained rules: %s (%s -> %s) | %s (%s -> %s)'\n % (action, tag, final_tags,\n action2, final_tag, final_tags2))\n if failchained and chainerror:\n raise ChainRuleError('Chained rules for tags detected!')\n Tags._wildcard_tags = list()\n for tag in Tags._tags_dict:\n if '*' in tag:\n Tags._wildcard_tags.append(tag)\n return Tags._tags_dict, Tags._wildcard_tags\n", "role": "user" }, { "content": "tagscleanupdicts", "role": "assistant" } ]
def x(identifier, configuration=None): user = User(configuration=configuration) result = user._load_from_hdx('user', identifier) if result: return user return None
read_from_hdx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(identifier, configuration=None):\n user = User(configuration=configuration)\n result = user._load_from_hdx('user', identifier)\n if result:\n return user\n return None\n", "role": "user" }, { "content": "read_from_hdx", "role": "assistant" } ]
def x(self): capacity = self.data.get('capacity') if capacity is not None: del self.data['capacity'] self._update_in_hdx('user', 'id') if capacity is not None: self.data['capacity'] = capacity
update_in_hdx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n capacity = self.data.get('capacity')\n if capacity is not None:\n del self.data['capacity']\n self._update_in_hdx('user', 'id')\n if capacity is not None:\n self.data['capacity'] = capacity\n", "role": "user" }, { "content": "update_in_hdx", "role": "assistant" } ]
def x(self): capacity = self.data.get('capacity') if capacity is not None: del self.data['capacity'] self._create_in_hdx('user', 'id', 'name') if capacity is not None: self.data['capacity'] = capacity
create_in_hdx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n capacity = self.data.get('capacity')\n if capacity is not None:\n del self.data['capacity']\n self._create_in_hdx('user', 'id', 'name')\n if capacity is not None:\n self.data['capacity'] = capacity\n", "role": "user" }, { "content": "create_in_hdx", "role": "assistant" } ]
def x(self, subject, text_body, html_body=None, sender=None, **kwargs): self.configuration.emailer().send([self.data['email']], subject, text_body, html_body=html_body, sender=sender, **kwargs)
email
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, subject, text_body, html_body=None, sender=None, **kwargs):\n self.configuration.emailer().send([self.data['email']], subject,\n text_body, html_body=html_body, sender=sender, **kwargs)\n", "role": "user" }, { "content": "email", "role": "assistant" } ]
def x(configuration=None, **kwargs): user = User(configuration=configuration) user['id'] = 'all users' result = user._write_to_hdx('list', kwargs, 'id') users = list() if result: for userdict in result: user = User(userdict, configuration=configuration) users.append(user) else: logger.debug(result) return users
get_all_users
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(configuration=None, **kwargs):\n user = User(configuration=configuration)\n user['id'] = 'all users'\n result = user._write_to_hdx('list', kwargs, 'id')\n users = list()\n if result:\n for userdict in result:\n user = User(userdict, configuration=configuration)\n users.append(user)\n else:\n logger.debug(result)\n return users\n", "role": "user" }, { "content": "get_all_users", "role": "assistant" } ]
def x(users, subject, text_body, html_body=None, sender=None, configuration =None, **kwargs): if not users: raise ValueError('No users supplied') recipients = list() for user in users: recipients.append(user.data['email']) if configuration is None: configuration = users[0].configuration configuration.emailer().send(recipients, subject, text_body, html_body= html_body, sender=sender, **kwargs)
email_users
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(users, subject, text_body, html_body=None, sender=None, configuration\n =None, **kwargs):\n if not users:\n raise ValueError('No users supplied')\n recipients = list()\n for user in users:\n recipients.append(user.data['email'])\n if configuration is None:\n configuration = users[0].configuration\n configuration.emailer().send(recipients, subject, text_body, html_body=\n html_body, sender=sender, **kwargs)\n", "role": "user" }, { "content": "email_users", "role": "assistant" } ]
def x(self, permission='read'): success, result = self._read_from_hdx('user', self.data['name'], 'id', self.actions()['listorgs'], permission=permission) organizations = list() if success: for organizationdict in result: organization = hdx.data.organization.Organization.read_from_hdx( organizationdict['id']) organizations.append(organization) return organizations
get_organizations
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, permission='read'):\n success, result = self._read_from_hdx('user', self.data['name'], 'id',\n self.actions()['listorgs'], permission=permission)\n organizations = list()\n if success:\n for organizationdict in result:\n organization = hdx.data.organization.Organization.read_from_hdx(\n organizationdict['id'])\n organizations.append(organization)\n return organizations\n", "role": "user" }, { "content": "get_organizations", "role": "assistant" } ]
def x(projectmainfn, **kwargs): site_url = Configuration._create(**kwargs) logger.info('--------------------------------------------------') logger.info('> Using HDX Python API Library %s' % Configuration.apiversion) logger.info('> HDX Site: %s' % site_url) UserAgent.user_agent = Configuration.read().user_agent projectmainfn()
facade
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(projectmainfn, **kwargs):\n site_url = Configuration._create(**kwargs)\n logger.info('--------------------------------------------------')\n logger.info('> Using HDX Python API Library %s' % Configuration.apiversion)\n logger.info('> HDX Site: %s' % site_url)\n UserAgent.user_agent = Configuration.read().user_agent\n projectmainfn()\n", "role": "user" }, { "content": "facade", "role": "assistant" } ]
def x(config_path=None): if config_path: config = LintConfig.load_from_file(config_path) click.echo('Using config from {0}'.format(config_path)) elif os.path.exists(DEFAULT_CONFIG_FILE): config = LintConfig.load_from_file(DEFAULT_CONFIG_FILE) click.echo('Using config from {0}'.format(DEFAULT_CONFIG_FILE)) else: config = LintConfig() return config
get_lint_config
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(config_path=None):\n if config_path:\n config = LintConfig.load_from_file(config_path)\n click.echo('Using config from {0}'.format(config_path))\n elif os.path.exists(DEFAULT_CONFIG_FILE):\n config = LintConfig.load_from_file(DEFAULT_CONFIG_FILE)\n click.echo('Using config from {0}'.format(DEFAULT_CONFIG_FILE))\n else:\n config = LintConfig()\n return config\n", "role": "user" }, { "content": "get_lint_config", "role": "assistant" } ]
def x(list_files, config, ignore, path): files = MarkdownFileFinder.find_files(path) if list_files: echo_files(files) lint_config = get_lint_config(config) lint_config.apply_on_csv_string(ignore, lint_config.disable_rule) linter = MarkdownLinter(lint_config) error_count = linter.lint_files(files) exit(error_count)
cli
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(list_files, config, ignore, path):\n files = MarkdownFileFinder.find_files(path)\n if list_files:\n echo_files(files)\n lint_config = get_lint_config(config)\n lint_config.apply_on_csv_string(ignore, lint_config.disable_rule)\n linter = MarkdownLinter(lint_config)\n error_count = linter.lint_files(files)\n exit(error_count)\n", "role": "user" }, { "content": "cli", "role": "assistant" } ]
def x(self, check_interval=300): while True: if args.config: config = config_file_parser.get_configuration(args.config) access_key_id = config['access-key-id'] secret_access_key = config['secret-access-key'] region = config['region'] else: access_key_id = args.access_key_id secret_access_key = args.secret_access_key region = args.region connection = connection_manager.connect_to_ec2(region, access_key_id, secret_access_key) snapshot_manager.run(connection) logger.info('Sleeping {} seconds until next check'.format( check_interval)) time.sleep(check_interval)
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, check_interval=300):\n while True:\n if args.config:\n config = config_file_parser.get_configuration(args.config)\n access_key_id = config['access-key-id']\n secret_access_key = config['secret-access-key']\n region = config['region']\n else:\n access_key_id = args.access_key_id\n secret_access_key = args.secret_access_key\n region = args.region\n connection = connection_manager.connect_to_ec2(region,\n access_key_id, secret_access_key)\n snapshot_manager.run(connection)\n logger.info('Sleeping {} seconds until next check'.format(\n check_interval))\n time.sleep(check_interval)\n", "role": "user" }, { "content": "run", "role": "assistant" } ]
def x(self, markdown_string): all_violations = [] lines = markdown_string.split('\n') line_rules = self.line_rules line_nr = 1 ignoring = False for line in lines: if ignoring: if line.strip() == '<!-- markdownlint:enable -->': ignoring = False else: if line.strip() == '<!-- markdownlint:disable -->': ignoring = True continue for rule in line_rules: violation = rule.validate(line) if violation: violation.line_nr = line_nr all_violations.append(violation) line_nr += 1 return all_violations
apply_line_rules
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, markdown_string):\n all_violations = []\n lines = markdown_string.split('\\n')\n line_rules = self.line_rules\n line_nr = 1\n ignoring = False\n for line in lines:\n if ignoring:\n if line.strip() == '<!-- markdownlint:enable -->':\n ignoring = False\n else:\n if line.strip() == '<!-- markdownlint:disable -->':\n ignoring = True\n continue\n for rule in line_rules:\n violation = rule.validate(line)\n if violation:\n violation.line_nr = line_nr\n all_violations.append(violation)\n line_nr += 1\n return all_violations\n", "role": "user" }, { "content": "apply_line_rules", "role": "assistant" } ]
def x(self, files): all_violations = [] for filename in files: with open(filename, 'r') as f: content = f.read() violations = self.lint(content) all_violations.extend(violations) for e in violations: print('{0}:{1}: {2} {3}'.format(filename, e.line_nr, e. rule_id, e.message)) return len(all_violations)
lint_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, files):\n all_violations = []\n for filename in files:\n with open(filename, 'r') as f:\n content = f.read()\n violations = self.lint(content)\n all_violations.extend(violations)\n for e in violations:\n print('{0}:{1}: {2} {3}'.format(filename, e.line_nr, e.\n rule_id, e.message))\n return len(all_violations)\n", "role": "user" }, { "content": "lint_files", "role": "assistant" } ]
def x(self, byte_stream): try: return self._struct.unpack_from(byte_stream) except (TypeError, struct.error) as exception: raise IOError('Unable to read byte stream with error: {0!s}'.format (exception))
ReadFrom
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, byte_stream):\n try:\n return self._struct.unpack_from(byte_stream)\n except (TypeError, struct.error) as exception:\n raise IOError('Unable to read byte stream with error: {0!s}'.format\n (exception))\n", "role": "user" }, { "content": "ReadFrom", "role": "assistant" } ]
def x(self, values): try: return self._struct.pack(*values) except (TypeError, struct.error) as exception: raise IOError('Unable to write stream with error: {0!s}'.format( exception))
WriteTo
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, values):\n try:\n return self._struct.pack(*values)\n except (TypeError, struct.error) as exception:\n raise IOError('Unable to write stream with error: {0!s}'.format(\n exception))\n", "role": "user" }, { "content": "WriteTo", "role": "assistant" } ]
def x(connection): volumes = volume_manager.get_watched_volumes(connection) for volume in volumes: _ensure_snapshot(connection, volume) _remove_old_snapshots(connection, volume)
run
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection):\n volumes = volume_manager.get_watched_volumes(connection)\n for volume in volumes:\n _ensure_snapshot(connection, volume)\n _remove_old_snapshots(connection, volume)\n", "role": "user" }, { "content": "run", "role": "assistant" } ]
def x(volume): logger.info('Creating new snapshot for {}'.format(volume.id)) snapshot = volume.create_snapshot(description= 'Automatic snapshot by Automated EBS Snapshots') logger.info('Created snapshot {} for volume {}'.format(snapshot.id, volume.id)) return snapshot
create_snapshot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(volume):\n logger.info('Creating new snapshot for {}'.format(volume.id))\n snapshot = volume.create_snapshot(description=\n 'Automatic snapshot by Automated EBS Snapshots')\n logger.info('Created snapshot {} for volume {}'.format(snapshot.id,\n volume.id))\n return snapshot\n", "role": "user" }, { "content": "create_snapshot", "role": "assistant" } ]
def x(connection, volume): if 'AutomatedEBSSnapshots' not in volume.tags: logger.warning('Missing tag AutomatedEBSSnapshots for volume {}'. format(volume.id)) return interval = volume.tags['AutomatedEBSSnapshots'] if volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS: logger.warning( '"{}" is not a valid snapshotting interval for volume {}'. format(interval, volume.id)) return snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id}) if not snapshots: _create_snapshot(volume) return min_delta = 3600 * 24 * 365 * 10 for snapshot in snapshots: timestamp = datetime.datetime.strptime(snapshot.start_time, '%Y-%m-%dT%H:%M:%S.000Z') delta_seconds = int((datetime.datetime.utcnow() - timestamp). total_seconds()) if delta_seconds < min_delta: min_delta = delta_seconds logger.info('The newest snapshot for {} is {} seconds old'.format( volume.id, min_delta)) if interval == 'hourly' and min_delta > 3600: _create_snapshot(volume) elif interval == 'daily' and min_delta > 3600 * 24: _create_snapshot(volume) elif interval == 'weekly' and min_delta > 3600 * 24 * 7: _create_snapshot(volume) elif interval == 'monthly' and min_delta > 3600 * 24 * 30: _create_snapshot(volume) elif interval == 'yearly' and min_delta > 3600 * 24 * 365: _create_snapshot(volume) else: logger.info('No need for a new snapshot of {}'.format(volume.id))
ensure_snapshot
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, volume):\n if 'AutomatedEBSSnapshots' not in volume.tags:\n logger.warning('Missing tag AutomatedEBSSnapshots for volume {}'.\n format(volume.id))\n return\n interval = volume.tags['AutomatedEBSSnapshots']\n if volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS:\n logger.warning(\n '\"{}\" is not a valid snapshotting interval for volume {}'.\n format(interval, volume.id))\n return\n snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id})\n if not snapshots:\n _create_snapshot(volume)\n return\n min_delta = 3600 * 24 * 365 * 10\n for snapshot in snapshots:\n timestamp = datetime.datetime.strptime(snapshot.start_time,\n '%Y-%m-%dT%H:%M:%S.000Z')\n delta_seconds = int((datetime.datetime.utcnow() - timestamp).\n total_seconds())\n if delta_seconds < min_delta:\n min_delta = delta_seconds\n logger.info('The newest snapshot for {} is {} seconds old'.format(\n volume.id, min_delta))\n if interval == 'hourly' and min_delta > 3600:\n _create_snapshot(volume)\n elif interval == 'daily' and min_delta > 3600 * 24:\n _create_snapshot(volume)\n elif interval == 'weekly' and min_delta > 3600 * 24 * 7:\n _create_snapshot(volume)\n elif interval == 'monthly' and min_delta > 3600 * 24 * 30:\n _create_snapshot(volume)\n elif interval == 'yearly' and min_delta > 3600 * 24 * 365:\n _create_snapshot(volume)\n else:\n logger.info('No need for a new snapshot of {}'.format(volume.id))\n", "role": "user" }, { "content": "ensure_snapshot", "role": "assistant" } ]
def x(connection, volume): if 'AutomatedEBSSnapshotsRetention' not in volume.tags: logger.warning( 'Missing tag AutomatedEBSSnapshotsRetention for volume {}'. format(volume.id)) return retention = int(volume.tags['AutomatedEBSSnapshotsRetention']) snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id}) snapshots.sort(key=lambda x: x.start_time) snapshots = snapshots[:-int(retention)] if not snapshots: logger.info('No old snapshots to remove') return for snapshot in snapshots: logger.info('Deleting snapshot {}'.format(snapshot.id)) try: snapshot.delete() except EC2ResponseError as error: logger.warning('Could not remove snapshot: {}'.format(error. message)) logger.info('Done deleting snapshots')
remove_old_snapshots
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, volume):\n if 'AutomatedEBSSnapshotsRetention' not in volume.tags:\n logger.warning(\n 'Missing tag AutomatedEBSSnapshotsRetention for volume {}'.\n format(volume.id))\n return\n retention = int(volume.tags['AutomatedEBSSnapshotsRetention'])\n snapshots = connection.get_all_snapshots(filters={'volume-id': volume.id})\n snapshots.sort(key=lambda x: x.start_time)\n snapshots = snapshots[:-int(retention)]\n if not snapshots:\n logger.info('No old snapshots to remove')\n return\n for snapshot in snapshots:\n logger.info('Deleting snapshot {}'.format(snapshot.id))\n try:\n snapshot.delete()\n except EC2ResponseError as error:\n logger.warning('Could not remove snapshot: {}'.format(error.\n message))\n logger.info('Done deleting snapshots')\n", "role": "user" }, { "content": "remove_old_snapshots", "role": "assistant" } ]
def x(connection): volumes = get_watched_volumes(connection) if not volumes: logger.info('No watched volumes found') return logger.info( '+-----------------------+----------------------+--------------+------------+' ) logger.info( '| {volume:<21} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |' .format(volume='Volume ID', volume_name='Volume name', interval= 'Interval', retention='Retention')) logger.info( '+-----------------------+----------------------+--------------+------------+' ) for volume in volumes: if 'AutomatedEBSSnapshots' not in volume.tags: interval = 'Interval tag not found' elif volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS: interval = 'Invalid interval' else: interval = volume.tags['AutomatedEBSSnapshots'] if 'AutomatedEBSSnapshotsRetention' not in volume.tags: retention = 0 else: retention = volume.tags['AutomatedEBSSnapshotsRetention'] try: volume_name = volume.tags['Name'] except KeyError: volume_name = '' logger.info( '|\xa0{volume_id:<14} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |' .format(volume_id=volume.id, volume_name=volume_name, interval= interval, retention=retention)) logger.info( '+-----------------------+----------------------+--------------+------------+' )
list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection):\n volumes = get_watched_volumes(connection)\n if not volumes:\n logger.info('No watched volumes found')\n return\n logger.info(\n '+-----------------------+----------------------+--------------+------------+'\n )\n logger.info(\n '| {volume:<21} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |'\n .format(volume='Volume ID', volume_name='Volume name', interval=\n 'Interval', retention='Retention'))\n logger.info(\n '+-----------------------+----------------------+--------------+------------+'\n )\n for volume in volumes:\n if 'AutomatedEBSSnapshots' not in volume.tags:\n interval = 'Interval tag not found'\n elif volume.tags['AutomatedEBSSnapshots'] not in VALID_INTERVALS:\n interval = 'Invalid interval'\n else:\n interval = volume.tags['AutomatedEBSSnapshots']\n if 'AutomatedEBSSnapshotsRetention' not in volume.tags:\n retention = 0\n else:\n retention = volume.tags['AutomatedEBSSnapshotsRetention']\n try:\n volume_name = volume.tags['Name']\n except KeyError:\n volume_name = ''\n logger.info(\n '|\\xa0{volume_id:<14} | {volume_name:<20.20} | {interval:<12} | {retention:<10} |'\n .format(volume_id=volume.id, volume_name=volume_name, interval=\n interval, retention=retention))\n logger.info(\n '+-----------------------+----------------------+--------------+------------+'\n )\n", "role": "user" }, { "content": "list", "role": "assistant" } ]
def x(connection, volume_id): try: volume = connection.get_all_volumes(volume_ids=[volume_id])[0] volume.remove_tag('AutomatedEBSSnapshots') except EC2ResponseError: pass logger.info('Removed {} from the watchlist'.format(volume_id)) return True
unwatch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, volume_id):\n try:\n volume = connection.get_all_volumes(volume_ids=[volume_id])[0]\n volume.remove_tag('AutomatedEBSSnapshots')\n except EC2ResponseError:\n pass\n logger.info('Removed {} from the watchlist'.format(volume_id))\n return True\n", "role": "user" }, { "content": "unwatch", "role": "assistant" } ]
def x(connection, volume_id, interval='daily', retention=0): try: volume = connection.get_all_volumes(volume_ids=[volume_id])[0] except EC2ResponseError: logger.warning('Volume {} not found'.format(volume_id)) return False if interval not in VALID_INTERVALS: logger.warning('{} is not a valid interval. Valid intervals are {}' .format(interval, ', '.join(VALID_INTERVALS))) volume.remove_tag('AutomatedEBSSnapshots') volume.add_tag('AutomatedEBSSnapshots', value=interval) volume.remove_tag('AutomatedEBSSnapshotsRetention') volume.add_tag('AutomatedEBSSnapshotsRetention', value=int(retention)) logger.info('Updated the rotation interval to {} for {}'.format( interval, volume_id)) return True
watch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, volume_id, interval='daily', retention=0):\n try:\n volume = connection.get_all_volumes(volume_ids=[volume_id])[0]\n except EC2ResponseError:\n logger.warning('Volume {} not found'.format(volume_id))\n return False\n if interval not in VALID_INTERVALS:\n logger.warning('{} is not a valid interval. Valid intervals are {}'\n .format(interval, ', '.join(VALID_INTERVALS)))\n volume.remove_tag('AutomatedEBSSnapshots')\n volume.add_tag('AutomatedEBSSnapshots', value=interval)\n volume.remove_tag('AutomatedEBSSnapshotsRetention')\n volume.add_tag('AutomatedEBSSnapshotsRetention', value=int(retention))\n logger.info('Updated the rotation interval to {} for {}'.format(\n interval, volume_id))\n return True\n", "role": "user" }, { "content": "watch", "role": "assistant" } ]
def x(connection, volume): volume_id_pattern = re.compile('vol-\\w{8}') if volume_id_pattern.match(volume): try: connection.get_all_volumes(volume_ids=[volume]) volume_id = volume except EC2ResponseError: logger.warning('Volume {} not found'.format(volume)) return None else: name_filter = {'tag-key': 'Name', 'tag-value': volume} volumes = connection.get_all_volumes(filters=name_filter) if not volumes: logger.warning('Volume {} not found'.format(volume)) return None if len(volumes) > 1: logger.warning('Volume {} not unique'.format(volume)) volume_id = volumes[0].id return volume_id
get_volume_id
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, volume):\n volume_id_pattern = re.compile('vol-\\\\w{8}')\n if volume_id_pattern.match(volume):\n try:\n connection.get_all_volumes(volume_ids=[volume])\n volume_id = volume\n except EC2ResponseError:\n logger.warning('Volume {} not found'.format(volume))\n return None\n else:\n name_filter = {'tag-key': 'Name', 'tag-value': volume}\n volumes = connection.get_all_volumes(filters=name_filter)\n if not volumes:\n logger.warning('Volume {} not found'.format(volume))\n return None\n if len(volumes) > 1:\n logger.warning('Volume {} not unique'.format(volume))\n volume_id = volumes[0].id\n return volume_id\n", "role": "user" }, { "content": "get_volume_id", "role": "assistant" } ]
def x(connection, file_name): with open(file_name, 'r') as filehandle: for line in filehandle.xreadlines(): volume, interval, retention = line.rstrip().split(',') watch(connection, get_volume_id(connection, volume), interval, retention)
watch_from_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, file_name):\n with open(file_name, 'r') as filehandle:\n for line in filehandle.xreadlines():\n volume, interval, retention = line.rstrip().split(',')\n watch(connection, get_volume_id(connection, volume), interval,\n retention)\n", "role": "user" }, { "content": "watch_from_file", "role": "assistant" } ]
def x(connection, file_name): with open(file_name, 'r') as filehandle: for line in filehandle.xreadlines(): volume, interval, retention = line.rstrip().split(',') unwatch(connection, get_volume_id(connection, volume))
unwatch_from_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, file_name):\n with open(file_name, 'r') as filehandle:\n for line in filehandle.xreadlines():\n volume, interval, retention = line.rstrip().split(',')\n unwatch(connection, get_volume_id(connection, volume))\n", "role": "user" }, { "content": "unwatch_from_file", "role": "assistant" } ]
def x(connection, volume): logger.info( '+----------------+----------------------+---------------------------+' ) logger.info('| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |' .format(snapshot='Snapshot ID', snapshot_name='Snapshot name', created='Created')) logger.info( '+----------------+----------------------+---------------------------+' ) vid = get_volume_id(connection, volume) if vid: vol = connection.get_all_volumes(volume_ids=[vid])[0] for snap in vol.snapshots(): logger.info( '| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |' .format(snapshot=snap.id, snapshot_name=snap.tags.get( 'Name', ''), created=snap.start_time)) logger.info( '+----------------+----------------------+---------------------------+' )
list_snapshots
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(connection, volume):\n logger.info(\n '+----------------+----------------------+---------------------------+'\n )\n logger.info('| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |'\n .format(snapshot='Snapshot ID', snapshot_name='Snapshot name',\n created='Created'))\n logger.info(\n '+----------------+----------------------+---------------------------+'\n )\n vid = get_volume_id(connection, volume)\n if vid:\n vol = connection.get_all_volumes(volume_ids=[vid])[0]\n for snap in vol.snapshots():\n logger.info(\n '| {snapshot:<14} | {snapshot_name:<20.20} | {created:<25} |'\n .format(snapshot=snap.id, snapshot_name=snap.tags.get(\n 'Name', ''), created=snap.start_time))\n logger.info(\n '+----------------+----------------------+---------------------------+'\n )\n", "role": "user" }, { "content": "list_snapshots", "role": "assistant" } ]
def x(self, words, parser, **kwargs): output = self._run_morfologik(words) return parser.parse(output, **kwargs)
stem
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, words, parser, **kwargs):\n output = self._run_morfologik(words)\n return parser.parse(output, **kwargs)\n", "role": "user" }, { "content": "stem", "role": "assistant" } ]
def x(self, words): p = subprocess.Popen(['java', '-jar', self.jar_path, 'plstem', '-ie', 'UTF-8', '-oe', 'UTF-8'], bufsize=-1, stdin=subprocess.PIPE, stdout =subprocess.PIPE, stderr=subprocess.STDOUT) out, _ = p.communicate(input=bytes('\n'.join(words), 'utf-8')) return decode(out, 'utf-8')
run_morfologik
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, words):\n p = subprocess.Popen(['java', '-jar', self.jar_path, 'plstem', '-ie',\n 'UTF-8', '-oe', 'UTF-8'], bufsize=-1, stdin=subprocess.PIPE, stdout\n =subprocess.PIPE, stderr=subprocess.STDOUT)\n out, _ = p.communicate(input=bytes('\\n'.join(words), 'utf-8'))\n return decode(out, 'utf-8')\n", "role": "user" }, { "content": "run_morfologik", "role": "assistant" } ]
def x(identifier, configuration=None): showcase = Showcase(configuration=configuration) result = showcase._load_from_hdx('showcase', identifier) if result: return showcase return None
read_from_hdx
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(identifier, configuration=None):\n showcase = Showcase(configuration=configuration)\n result = showcase._load_from_hdx('showcase', identifier)\n if result:\n return showcase\n return None\n", "role": "user" }, { "content": "read_from_hdx", "role": "assistant" } ]
def x(self): assoc_result, datasets_dicts = self._read_from_hdx('showcase', self. data['id'], fieldname='showcase_id', action=self.actions()[ 'list_datasets']) datasets = list() if assoc_result: for dataset_dict in datasets_dicts: dataset = hdx.data.dataset.Dataset(dataset_dict, configuration= self.configuration) datasets.append(dataset) return datasets
get_datasets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n assoc_result, datasets_dicts = self._read_from_hdx('showcase', self.\n data['id'], fieldname='showcase_id', action=self.actions()[\n 'list_datasets'])\n datasets = list()\n if assoc_result:\n for dataset_dict in datasets_dicts:\n dataset = hdx.data.dataset.Dataset(dataset_dict, configuration=\n self.configuration)\n datasets.append(dataset)\n return datasets\n", "role": "user" }, { "content": "get_datasets", "role": "assistant" } ]
def x(self, dataset): if isinstance(dataset, hdx.data.dataset.Dataset) or isinstance(dataset, dict): if 'id' not in dataset: dataset = hdx.data.dataset.Dataset.read_from_hdx(dataset['name']) dataset = dataset['id'] elif not isinstance(dataset, str): raise hdx.data.hdxobject.HDXError( 'Type %s cannot be added as a dataset!' % type(dataset).__name__) if is_valid_uuid(dataset) is False: raise hdx.data.hdxobject.HDXError('%s is not a valid dataset id!' % dataset) return {'showcase_id': self.data['id'], 'package_id': dataset}
get_showcase_dataset_dict
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dataset):\n if isinstance(dataset, hdx.data.dataset.Dataset) or isinstance(dataset,\n dict):\n if 'id' not in dataset:\n dataset = hdx.data.dataset.Dataset.read_from_hdx(dataset['name'])\n dataset = dataset['id']\n elif not isinstance(dataset, str):\n raise hdx.data.hdxobject.HDXError(\n 'Type %s cannot be added as a dataset!' % type(dataset).__name__)\n if is_valid_uuid(dataset) is False:\n raise hdx.data.hdxobject.HDXError('%s is not a valid dataset id!' %\n dataset)\n return {'showcase_id': self.data['id'], 'package_id': dataset}\n", "role": "user" }, { "content": "get_showcase_dataset_dict", "role": "assistant" } ]
def x(self, dataset, datasets_to_check=None): showcase_dataset = self._get_showcase_dataset_dict(dataset) if datasets_to_check is None: datasets_to_check = self.get_datasets() for dataset in datasets_to_check: if showcase_dataset['package_id'] == dataset['id']: return False self._write_to_hdx('associate', showcase_dataset, 'package_id') return True
add_dataset
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, dataset, datasets_to_check=None):\n showcase_dataset = self._get_showcase_dataset_dict(dataset)\n if datasets_to_check is None:\n datasets_to_check = self.get_datasets()\n for dataset in datasets_to_check:\n if showcase_dataset['package_id'] == dataset['id']:\n return False\n self._write_to_hdx('associate', showcase_dataset, 'package_id')\n return True\n", "role": "user" }, { "content": "add_dataset", "role": "assistant" } ]
def x(self, datasets, datasets_to_check=None): if datasets_to_check is None: datasets_to_check = self.get_datasets() alldatasetsadded = True for dataset in datasets: if not self.add_dataset(dataset, datasets_to_check=datasets_to_check): alldatasetsadded = False return alldatasetsadded
add_datasets
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, datasets, datasets_to_check=None):\n if datasets_to_check is None:\n datasets_to_check = self.get_datasets()\n alldatasetsadded = True\n for dataset in datasets:\n if not self.add_dataset(dataset, datasets_to_check=datasets_to_check):\n alldatasetsadded = False\n return alldatasetsadded\n", "role": "user" }, { "content": "add_datasets", "role": "assistant" } ]
def x(nb, jsonable_parameter=True, end_cell_index=None): jh = _JupyterNotebookHelper(nb, jsonable_parameter, end_cell_index) if jsonable_parameter is True: PossibleParameter = collections.namedtuple('PossibleParameter', [ 'name', 'value', 'cell_index']) else: PossibleParameter = collections.namedtuple('PossibleParameter', [ 'name', 'cell_index']) res = [] for name, cell_index in jh.param_cell_index.items(): if jsonable_parameter is True: res.append(PossibleParameter(name=name, value=jh.param_value[ name], cell_index=cell_index)) else: res.append(PossibleParameter(name=name, cell_index=cell_index)) return sorted(res, key=lambda x: x.name)
possible_parameter
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(nb, jsonable_parameter=True, end_cell_index=None):\n jh = _JupyterNotebookHelper(nb, jsonable_parameter, end_cell_index)\n if jsonable_parameter is True:\n PossibleParameter = collections.namedtuple('PossibleParameter', [\n 'name', 'value', 'cell_index'])\n else:\n PossibleParameter = collections.namedtuple('PossibleParameter', [\n 'name', 'cell_index'])\n res = []\n for name, cell_index in jh.param_cell_index.items():\n if jsonable_parameter is True:\n res.append(PossibleParameter(name=name, value=jh.param_value[\n name], cell_index=cell_index))\n else:\n res.append(PossibleParameter(name=name, cell_index=cell_index))\n return sorted(res, key=lambda x: x.name)\n", "role": "user" }, { "content": "possible_parameter", "role": "assistant" } ]
def x(self, joiner, formatter=lambda s, t: t.format(s), template='{}'): return ww.s(joiner).join(self, formatter, template)
join
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, joiner, formatter=lambda s, t: t.format(s), template='{}'):\n return ww.s(joiner).join(self, formatter, template)\n", "role": "user" }, { "content": "join", "role": "assistant" } ]
def x(self, *values): for value in values: list.append(self, value) return self
append
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *values):\n for value in values:\n list.append(self, value)\n return self\n", "role": "user" }, { "content": "append", "role": "assistant" } ]
def x(self, *iterables): for value in iterables: list.extend(self, value) return self
extend
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, *iterables):\n for value in iterables:\n list.extend(self, value)\n return self\n", "role": "user" }, { "content": "extend", "role": "assistant" } ]
def x(value): if isinstance(value, dict) or isinstance(value, list): return json.dumps(value) return value
normalize_cell_value
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(value):\n if isinstance(value, dict) or isinstance(value, list):\n return json.dumps(value)\n return value\n", "role": "user" }, { "content": "normalize_cell_value", "role": "assistant" } ]
def x(input_file_name): mode = 'r' if sys.version_info[0] < 3: mode = 'rb' with io.open(input_file_name, mode) as input_file: reader = csv.reader(input_file, delimiter=',', quotechar='"') addresses = list(map(tuple, reader)) if len(addresses) == 0: raise Exception('No addresses found in input file') header_columns = list(column.lower() for column in addresses.pop(0)) try: address_index = header_columns.index('address') zipcode_index = header_columns.index('zipcode') except ValueError: raise Exception( "The first row of the input CSV must be a header that contains a column labeled 'address' and a column labeled 'zipcode'." ) return list((row[address_index], row[zipcode_index]) for row in addresses)
get_addresses_from_input_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_file_name):\n mode = 'r'\n if sys.version_info[0] < 3:\n mode = 'rb'\n with io.open(input_file_name, mode) as input_file:\n reader = csv.reader(input_file, delimiter=',', quotechar='\"')\n addresses = list(map(tuple, reader))\n if len(addresses) == 0:\n raise Exception('No addresses found in input file')\n header_columns = list(column.lower() for column in addresses.pop(0))\n try:\n address_index = header_columns.index('address')\n zipcode_index = header_columns.index('zipcode')\n except ValueError:\n raise Exception(\n \"The first row of the input CSV must be a header that contains a column labeled 'address' and a column labeled 'zipcode'.\"\n )\n return list((row[address_index], row[zipcode_index]) for row in\n addresses)\n", "role": "user" }, { "content": "get_addresses_from_input_file", "role": "assistant" } ]
def x(input_file_name): valid_identifiers = ['address', 'zipcode', 'unit', 'city', 'state', 'slug', 'block_id', 'msa', 'num_bins', 'property_type', 'client_value', 'client_value_sqft', 'meta'] mode = 'r' if sys.version_info[0] < 3: mode = 'rb' with io.open(input_file_name, mode) as input_file: result = [{identifier: val for identifier, val in list(row.items()) if identifier in valid_identifiers} for row in csv.DictReader( input_file, skipinitialspace=True)] return result
get_identifiers_from_input_file
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(input_file_name):\n valid_identifiers = ['address', 'zipcode', 'unit', 'city', 'state',\n 'slug', 'block_id', 'msa', 'num_bins', 'property_type',\n 'client_value', 'client_value_sqft', 'meta']\n mode = 'r'\n if sys.version_info[0] < 3:\n mode = 'rb'\n with io.open(input_file_name, mode) as input_file:\n result = [{identifier: val for identifier, val in list(row.items()) if\n identifier in valid_identifiers} for row in csv.DictReader(\n input_file, skipinitialspace=True)]\n return result\n", "role": "user" }, { "content": "get_identifiers_from_input_file", "role": "assistant" } ]
def x(ipAddress, auth, url): real_time_locate_url = ( '/imcrs/res/access/realtimeLocate?type=2&value=' + str(ipAddress) + '&total=false') f_url = url + real_time_locate_url r = requests.get(f_url, auth=auth, headers=HEADERS) try: if r.status_code == 200: response = json.loads(r.text) if 'realtimeLocation' in response: real_time_locate = json.loads(r.text)['realtimeLocation'] if type(real_time_locate) is dict: real_time_locate = [real_time_locate] return real_time_locate else: return json.loads(r.text)['realtimeLocation'] else: return json.loads(r.text) except requests.exceptions.RequestException as e: return 'Error:\n' + str(e ) + ' get_real_time_locate: An Error has occured'
get_real_time_locate
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ipAddress, auth, url):\n real_time_locate_url = (\n '/imcrs/res/access/realtimeLocate?type=2&value=' + str(ipAddress) +\n '&total=false')\n f_url = url + real_time_locate_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n response = json.loads(r.text)\n if 'realtimeLocation' in response:\n real_time_locate = json.loads(r.text)['realtimeLocation']\n if type(real_time_locate) is dict:\n real_time_locate = [real_time_locate]\n return real_time_locate\n else:\n return json.loads(r.text)['realtimeLocation']\n else:\n return json.loads(r.text)\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e\n ) + ' get_real_time_locate: An Error has occured'\n", "role": "user" }, { "content": "get_real_time_locate", "role": "assistant" } ]
def x(devId, auth, url): if auth is None or url is None: set_imc_creds() ip_mac_arp_list_url = '/imcrs/res/access/ipMacArp/' + str(devId) f_url = url + ip_mac_arp_list_url r = requests.get(f_url, auth=auth, headers=HEADERS) try: if r.status_code == 200: macarplist = json.loads(r.text) if len(macarplist) > 1: return macarplist['ipMacArp'] else: return ['this function is unsupported'] except requests.exceptions.RequestException as e: return 'Error:\n' + str(e ) + ' get_ip_mac_arp_list: An Error has occured'
get_ip_mac_arp_list
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(devId, auth, url):\n if auth is None or url is None:\n set_imc_creds()\n ip_mac_arp_list_url = '/imcrs/res/access/ipMacArp/' + str(devId)\n f_url = url + ip_mac_arp_list_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n macarplist = json.loads(r.text)\n if len(macarplist) > 1:\n return macarplist['ipMacArp']\n else:\n return ['this function is unsupported']\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e\n ) + ' get_ip_mac_arp_list: An Error has occured'\n", "role": "user" }, { "content": "get_ip_mac_arp_list", "role": "assistant" } ]
def x(scopeId, auth, url): get_ip_scope_url = '/imcrs/res/access/assignedIpScope/' + str(scopeId) f_url = url + get_ip_scope_url r = requests.get(f_url, auth=auth, headers=HEADERS) try: if r.status_code == 200: ipscopelist = json.loads(r.text) return ipscopelist except requests.exceptions.RequestException as e: return 'Error:\n' + str(e) + ' get_ip_scope: An Error has occured'
get_ip_scope_detail
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(scopeId, auth, url):\n get_ip_scope_url = '/imcrs/res/access/assignedIpScope/' + str(scopeId)\n f_url = url + get_ip_scope_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n ipscopelist = json.loads(r.text)\n return ipscopelist\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' get_ip_scope: An Error has occured'\n", "role": "user" }, { "content": "get_ip_scope_detail", "role": "assistant" } ]
def x(network_address, auth, url): scope_id = get_scope_id(network_address, auth, url) delete_ip_address_url = '/imcrs/res/access/assignedIpScope/' + str(scope_id ) f_url = url + delete_ip_address_url r = requests.delete(f_url, auth=auth, headers=HEADERS) try: return r if r.status_code == 204: return r.status_code except requests.exceptions.RequestException as e: return 'Error:\n' + str(e) + ' delete_ip_scope: An Error has occured'
delete_ip_scope
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(network_address, auth, url):\n scope_id = get_scope_id(network_address, auth, url)\n delete_ip_address_url = '/imcrs/res/access/assignedIpScope/' + str(scope_id\n )\n f_url = url + delete_ip_address_url\n r = requests.delete(f_url, auth=auth, headers=HEADERS)\n try:\n return r\n if r.status_code == 204:\n return r.status_code\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' delete_ip_scope: An Error has occured'\n", "role": "user" }, { "content": "delete_ip_scope", "role": "assistant" } ]
def x(ipaddress, name, description, scopeid, auth, url): new_ip = {'ip': ipaddress, 'name': name, 'description': description} add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip?ipScopeId=' + str( scopeid) f_url = url + add_scope_ip_url payload = json.dumps(new_ip) r = requests.post(f_url, auth=auth, headers=HEADERS, data=payload) try: if r.status_code == 200: return r.status_code elif r.status_code == 409: return r.status_code except requests.exceptions.RequestException as e: return 'Error:\n' + str(e) + ' add_ip_scope: An Error has occured'
add_scope_ip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ipaddress, name, description, scopeid, auth, url):\n new_ip = {'ip': ipaddress, 'name': name, 'description': description}\n add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip?ipScopeId=' + str(\n scopeid)\n f_url = url + add_scope_ip_url\n payload = json.dumps(new_ip)\n r = requests.post(f_url, auth=auth, headers=HEADERS, data=payload)\n try:\n if r.status_code == 200:\n return r.status_code\n elif r.status_code == 409:\n return r.status_code\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' add_ip_scope: An Error has occured'\n", "role": "user" }, { "content": "add_scope_ip", "role": "assistant" } ]
def x(hostid, auth, url): add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip/' + str(hostid) f_url = url + add_scope_ip_url r = requests.delete(f_url, auth=auth, headers=HEADERS) try: if r.status_code == 204: return r.status_code elif r.status_code == 409: return r.status_code except requests.exceptions.RequestException as e: return 'Error:\n' + str(e) + ' add_ip_scope: An Error has occured'
remove_scope_ip
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(hostid, auth, url):\n add_scope_ip_url = '/imcrs/res/access/assignedIpScope/ip/' + str(hostid)\n f_url = url + add_scope_ip_url\n r = requests.delete(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 204:\n return r.status_code\n elif r.status_code == 409:\n return r.status_code\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' add_ip_scope: An Error has occured'\n", "role": "user" }, { "content": "remove_scope_ip", "role": "assistant" } ]
def x(scopeId, auth, url): get_ip_scope_url = ( '/imcrs/res/access/assignedIpScope/ip?size=10000&ipScopeId=' + str( scopeId)) f_url = url + get_ip_scope_url r = requests.get(f_url, auth=auth, headers=HEADERS) try: if r.status_code == 200: ipscopelist = json.loads(r.text) if ipscopelist == {}: return ipscopelist else: ipscopelist = ipscopelist['assignedIpInfo'] if type(ipscopelist) is dict: ipscope = [] ipscope.append(ipscopelist) return ipscope return ipscopelist except requests.exceptions.RequestException as e: return 'Error:\n' + str(e) + ' get_ip_scope: An Error has occured'
get_ip_scope_hosts
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(scopeId, auth, url):\n get_ip_scope_url = (\n '/imcrs/res/access/assignedIpScope/ip?size=10000&ipScopeId=' + str(\n scopeId))\n f_url = url + get_ip_scope_url\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n ipscopelist = json.loads(r.text)\n if ipscopelist == {}:\n return ipscopelist\n else:\n ipscopelist = ipscopelist['assignedIpInfo']\n if type(ipscopelist) is dict:\n ipscope = []\n ipscope.append(ipscopelist)\n return ipscope\n return ipscopelist\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' get_ip_scope: An Error has occured'\n", "role": "user" }, { "content": "get_ip_scope_hosts", "role": "assistant" } ]
def x(ipaddress, name, description, network_address, auth, url): scope_id = get_scope_id(network_address, auth, url) add_scope_ip(ipaddress, name, description, scope_id, auth, url)
add_host_to_segment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ipaddress, name, description, network_address, auth, url):\n scope_id = get_scope_id(network_address, auth, url)\n add_scope_ip(ipaddress, name, description, scope_id, auth, url)\n", "role": "user" }, { "content": "add_host_to_segment", "role": "assistant" } ]
def x(ipaddress, networkaddress, auth, url): host_id = get_host_id(ipaddress, networkaddress, auth, url) remove_scope_ip(host_id, auth.creds, auth.url)
delete_host_from_segment
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ipaddress, networkaddress, auth, url):\n host_id = get_host_id(ipaddress, networkaddress, auth, url)\n remove_scope_ip(host_id, auth.creds, auth.url)\n", "role": "user" }, { "content": "delete_host_from_segment", "role": "assistant" } ]
def x(method, version, endpoint, date, rel_url, content_type, content, access_key, secret_key, hash_type): hash_type = hash_type hostname = endpoint._val.netloc if version >= 'v4.20181215': content = b'' elif content_type.startswith('multipart/'): content = b'' body_hash = hashlib.new(hash_type, content).hexdigest() sign_str = ( '{}\n{}\n{}\nhost:{}\ncontent-type:{}\nx-backendai-version:{}\n{}'. format(method.upper(), rel_url, date.isoformat(), hostname, content_type.lower(), version, body_hash)) sign_bytes = sign_str.encode() sign_key = hmac.new(secret_key.encode(), date.strftime('%Y%m%d').encode (), hash_type).digest() sign_key = hmac.new(sign_key, hostname.encode(), hash_type).digest() signature = hmac.new(sign_key, sign_bytes, hash_type).hexdigest() headers = {'Authorization': 'BackendAI signMethod=HMAC-{}, credential={}:{}'.format(hash_type. upper(), access_key, signature)} return headers, signature
generate_signature
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(method, version, endpoint, date, rel_url, content_type, content,\n access_key, secret_key, hash_type):\n hash_type = hash_type\n hostname = endpoint._val.netloc\n if version >= 'v4.20181215':\n content = b''\n elif content_type.startswith('multipart/'):\n content = b''\n body_hash = hashlib.new(hash_type, content).hexdigest()\n sign_str = (\n '{}\\n{}\\n{}\\nhost:{}\\ncontent-type:{}\\nx-backendai-version:{}\\n{}'.\n format(method.upper(), rel_url, date.isoformat(), hostname,\n content_type.lower(), version, body_hash))\n sign_bytes = sign_str.encode()\n sign_key = hmac.new(secret_key.encode(), date.strftime('%Y%m%d').encode\n (), hash_type).digest()\n sign_key = hmac.new(sign_key, hostname.encode(), hash_type).digest()\n signature = hmac.new(sign_key, sign_bytes, hash_type).hexdigest()\n headers = {'Authorization':\n 'BackendAI signMethod=HMAC-{}, credential={}:{}'.format(hash_type.\n upper(), access_key, signature)}\n return headers, signature\n", "role": "user" }, { "content": "generate_signature", "role": "assistant" } ]
async def list_with_limit(cls, limit, offset, status: str='ALIVE', fields: Iterable[str]=None) ->Sequence[dict]: """ Fetches the list of agents with the given status with limit and offset for pagination. :param limit: number of agents to get :param offset: offset index of agents to get :param status: An upper-cased string constant representing agent status (one of ``'ALIVE'``, ``'TERMINATED'``, ``'LOST'``, etc.) :param fields: Additional per-agent query fields to fetch. """ if fields is None: fields = ('id', 'addr', 'status', 'first_contact', 'mem_slots', 'cpu_slots', 'gpu_slots') q = ( 'query($limit: Int!, $offset: Int!, $status: String) { agent_list(limit: $limit, offset: $offset, status: $status) { items { $fields } total_count }}' ) q = q.replace('$fields', ' '.join(fields)) variables = {'limit': limit, 'offset': offset, 'status': status} rqst = Request(cls.session, 'POST', '/admin/graphql') rqst.set_json({'query': q, 'variables': variables}) async with rqst.fetch() as resp: data = await resp.json() return data['agent_list']
list_with_limit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "async def list_with_limit(cls, limit, offset, status: str='ALIVE', fields:\n Iterable[str]=None) ->Sequence[dict]:\n \"\"\"\n Fetches the list of agents with the given status with limit and offset for\n pagination.\n\n :param limit: number of agents to get\n :param offset: offset index of agents to get\n :param status: An upper-cased string constant representing agent\n status (one of ``'ALIVE'``, ``'TERMINATED'``, ``'LOST'``,\n etc.)\n :param fields: Additional per-agent query fields to fetch.\n \"\"\"\n if fields is None:\n fields = ('id', 'addr', 'status', 'first_contact', 'mem_slots',\n 'cpu_slots', 'gpu_slots')\n q = (\n 'query($limit: Int!, $offset: Int!, $status: String) { agent_list(limit: $limit, offset: $offset, status: $status) { items { $fields } total_count }}'\n )\n q = q.replace('$fields', ' '.join(fields))\n variables = {'limit': limit, 'offset': offset, 'status': status}\n rqst = Request(cls.session, 'POST', '/admin/graphql')\n rqst.set_json({'query': q, 'variables': variables})\n async with rqst.fetch() as resp:\n data = await resp.json()\n return data['agent_list']\n", "role": "user" }, { "content": "list_with_limit", "role": "assistant" } ]
def x(self, value: RequestContent, *, content_type: str=None): assert self._attached_files is None, 'cannot set content because you already attached files.' guessed_content_type = 'application/octet-stream' if value is None: guessed_content_type = 'text/plain' self._content = b'' elif isinstance(value, str): guessed_content_type = 'text/plain' self._content = value.encode('utf-8') else: guessed_content_type = 'application/octet-stream' self._content = value self.content_type = (content_type if content_type is not None else guessed_content_type)
set_content
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value: RequestContent, *, content_type: str=None):\n assert self._attached_files is None, 'cannot set content because you already attached files.'\n guessed_content_type = 'application/octet-stream'\n if value is None:\n guessed_content_type = 'text/plain'\n self._content = b''\n elif isinstance(value, str):\n guessed_content_type = 'text/plain'\n self._content = value.encode('utf-8')\n else:\n guessed_content_type = 'application/octet-stream'\n self._content = value\n self.content_type = (content_type if content_type is not None else\n guessed_content_type)\n", "role": "user" }, { "content": "set_content", "role": "assistant" } ]
def x(self, value: object): self.set_content(modjson.dumps(value, cls=ExtendedJSONEncoder), content_type='application/json')
set_json
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, value: object):\n self.set_content(modjson.dumps(value, cls=ExtendedJSONEncoder),\n content_type='application/json')\n", "role": "user" }, { "content": "set_json", "role": "assistant" } ]
def x(self, files: Sequence[AttachedFile]): assert not self._content, 'content must be empty to attach files.' self.content_type = 'multipart/form-data' self._attached_files = files
attach_files
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, files: Sequence[AttachedFile]):\n assert not self._content, 'content must be empty to attach files.'\n self.content_type = 'multipart/form-data'\n self._attached_files = files\n", "role": "user" }, { "content": "attach_files", "role": "assistant" } ]
def x(self, rel_url, access_key=None, secret_key=None, hash_type=None): if access_key is None: access_key = self.config.access_key if secret_key is None: secret_key = self.config.secret_key if hash_type is None: hash_type = self.config.hash_type hdrs, _ = generate_signature(self.method, self.config.version, self. config.endpoint, self.date, str(rel_url), self.content_type, self. _content, access_key, secret_key, hash_type) self.headers.update(hdrs)
sign
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, rel_url, access_key=None, secret_key=None, hash_type=None):\n if access_key is None:\n access_key = self.config.access_key\n if secret_key is None:\n secret_key = self.config.secret_key\n if hash_type is None:\n hash_type = self.config.hash_type\n hdrs, _ = generate_signature(self.method, self.config.version, self.\n config.endpoint, self.date, str(rel_url), self.content_type, self.\n _content, access_key, secret_key, hash_type)\n self.headers.update(hdrs)\n", "role": "user" }, { "content": "sign", "role": "assistant" } ]
def x(self, **kwargs) ->'FetchContextManager': assert self.method in self._allowed_methods, 'Disallowed HTTP method: {}'.format( self.method) self.date = datetime.now(tzutc()) self.headers['Date'] = self.date.isoformat() if self.content_type is not None: self.headers['Content-Type'] = self.content_type full_url = self._build_url() self._sign(full_url.relative()) rqst_ctx = self.session.aiohttp_session.request(self.method, str( full_url), data=self._pack_content(), timeout= _default_request_timeout, headers=self.headers) return FetchContextManager(self.session, rqst_ctx, **kwargs)
fetch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs) ->'FetchContextManager':\n assert self.method in self._allowed_methods, 'Disallowed HTTP method: {}'.format(\n self.method)\n self.date = datetime.now(tzutc())\n self.headers['Date'] = self.date.isoformat()\n if self.content_type is not None:\n self.headers['Content-Type'] = self.content_type\n full_url = self._build_url()\n self._sign(full_url.relative())\n rqst_ctx = self.session.aiohttp_session.request(self.method, str(\n full_url), data=self._pack_content(), timeout=\n _default_request_timeout, headers=self.headers)\n return FetchContextManager(self.session, rqst_ctx, **kwargs)\n", "role": "user" }, { "content": "fetch", "role": "assistant" } ]
def x(self, **kwargs) ->'WebSocketContextManager': assert isinstance(self.session, AsyncSession ), 'Cannot use websockets with sessions in the synchronous mode' assert self.method == 'GET', 'Invalid websocket method' self.date = datetime.now(tzutc()) self.headers['Date'] = self.date.isoformat() self.content_type = 'application/octet-stream' full_url = self._build_url() self._sign(full_url.relative()) ws_ctx = self.session.aiohttp_session.ws_connect(str(full_url), autoping=True, heartbeat=30.0, headers=self.headers) return WebSocketContextManager(self.session, ws_ctx, **kwargs)
connect_websocket
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, **kwargs) ->'WebSocketContextManager':\n assert isinstance(self.session, AsyncSession\n ), 'Cannot use websockets with sessions in the synchronous mode'\n assert self.method == 'GET', 'Invalid websocket method'\n self.date = datetime.now(tzutc())\n self.headers['Date'] = self.date.isoformat()\n self.content_type = 'application/octet-stream'\n full_url = self._build_url()\n self._sign(full_url.relative())\n ws_ctx = self.session.aiohttp_session.ws_connect(str(full_url),\n autoping=True, heartbeat=30.0, headers=self.headers)\n return WebSocketContextManager(self.session, ws_ctx, **kwargs)\n", "role": "user" }, { "content": "connect_websocket", "role": "assistant" } ]
def x(): try: src_path = os.path.dirname(os.path.dirname(os.path.dirname(os.path. abspath(__file__)))) if src_path not in sys.path: sys.path.append(src_path) from aoiklivereload import LiveReloader reloader = LiveReloader(reload_mode='spawn_exit' if sys.platform == 'win32' else 'exec', force_exit=True) reloader.start_watcher_thread() server_host = '0.0.0.0' server_port = 8000 msg = '# ----- Run server -----\nHost: {}\nPort: {}'.format(server_host , server_port) print(msg) class HelloHandler(tornado.web.RequestHandler): """ Request handler class. """ def x(self): self.write('hello') handler_tuples = [('/', HelloHandler)] tornado_app = tornado.web.Application(handler_tuples, debug=False) tornado_app.listen(server_port, address=server_host) io_loop = tornado.ioloop.IOLoop.current() io_loop.start() except KeyboardInterrupt: pass
main
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x():\n try:\n src_path = os.path.dirname(os.path.dirname(os.path.dirname(os.path.\n abspath(__file__))))\n if src_path not in sys.path:\n sys.path.append(src_path)\n from aoiklivereload import LiveReloader\n reloader = LiveReloader(reload_mode='spawn_exit' if sys.platform ==\n 'win32' else 'exec', force_exit=True)\n reloader.start_watcher_thread()\n server_host = '0.0.0.0'\n server_port = 8000\n msg = '# ----- Run server -----\\nHost: {}\\nPort: {}'.format(server_host\n , server_port)\n print(msg)\n\n\n class HelloHandler(tornado.web.RequestHandler):\n \"\"\"\n Request handler class.\n \"\"\"\n\n def x(self):\n self.write('hello')\n handler_tuples = [('/', HelloHandler)]\n tornado_app = tornado.web.Application(handler_tuples, debug=False)\n tornado_app.listen(server_port, address=server_host)\n io_loop = tornado.ioloop.IOLoop.current()\n io_loop.start()\n except KeyboardInterrupt:\n pass\n", "role": "user" }, { "content": "main", "role": "assistant" } ]
def x(task, auth, url): add_perf_task_url = '/imcrs/perf/task' f_url = url + add_perf_task_url payload = json.dumps(task) response = requests.post(f_url, data=payload, auth=auth, headers=HEADERS) try: return response.status_code except requests.exceptions.RequestException as error: return 'Error:\n' + str(error) + ' add_perf_task: An Error has occured'
add_perf_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(task, auth, url):\n add_perf_task_url = '/imcrs/perf/task'\n f_url = url + add_perf_task_url\n payload = json.dumps(task)\n response = requests.post(f_url, data=payload, auth=auth, headers=HEADERS)\n try:\n return response.status_code\n except requests.exceptions.RequestException as error:\n return 'Error:\\n' + str(error) + ' add_perf_task: An Error has occured'\n", "role": "user" }, { "content": "add_perf_task", "role": "assistant" } ]
def x(task_name, auth, url): get_perf_task_url = ('/imcrs/perf/task?name=' + task_name + '&orderBy=taskId&desc=false') f_url = url + get_perf_task_url response = requests.get(f_url, auth=auth, headers=HEADERS) try: if response.status_code == 200: perf_task_info = json.loads(response.text) if 'task' in perf_task_info: perf_task_info = json.loads(response.text)['task'] else: perf_task_info = "Task Doesn't Exist" return perf_task_info except requests.exceptions.RequestException as error: return 'Error:\n' + str(error) + ' get_perf_task: An Error has occured'
get_perf_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(task_name, auth, url):\n get_perf_task_url = ('/imcrs/perf/task?name=' + task_name +\n '&orderBy=taskId&desc=false')\n f_url = url + get_perf_task_url\n response = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if response.status_code == 200:\n perf_task_info = json.loads(response.text)\n if 'task' in perf_task_info:\n perf_task_info = json.loads(response.text)['task']\n else:\n perf_task_info = \"Task Doesn't Exist\"\n return perf_task_info\n except requests.exceptions.RequestException as error:\n return 'Error:\\n' + str(error) + ' get_perf_task: An Error has occured'\n", "role": "user" }, { "content": "get_perf_task", "role": "assistant" } ]
def x(task_name, auth, url): task_id = get_perf_task(task_name, auth, url) if isinstance(task_id, str): print("Perf task doesn't exist") return 403 task_id = task_id['taskId'] get_perf_task_url = '/imcrs/perf/task/delete/' + str(task_id) f_url = url + get_perf_task_url response = requests.delete(f_url, auth=auth, headers=HEADERS) try: if response.status_code == 204: print('Perf Task successfully delete') return response.status_code except requests.exceptions.RequestException as error: return 'Error:\n' + str(error ) + ' delete_perf_task: An Error has occured'
delete_perf_task
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(task_name, auth, url):\n task_id = get_perf_task(task_name, auth, url)\n if isinstance(task_id, str):\n print(\"Perf task doesn't exist\")\n return 403\n task_id = task_id['taskId']\n get_perf_task_url = '/imcrs/perf/task/delete/' + str(task_id)\n f_url = url + get_perf_task_url\n response = requests.delete(f_url, auth=auth, headers=HEADERS)\n try:\n if response.status_code == 204:\n print('Perf Task successfully delete')\n return response.status_code\n except requests.exceptions.RequestException as error:\n return 'Error:\\n' + str(error\n ) + ' delete_perf_task: An Error has occured'\n", "role": "user" }, { "content": "delete_perf_task", "role": "assistant" } ]
def x(self, response): response_json = response.json() code_key = 'code' if code_key in response_json and response_json[code_key ] != constants.HTTP_CODE_OK: code = response_json[code_key] message = response_json if 'message' in response_json: message = response_json['message'] elif 'code_description' in response_json: message = response_json['code_description'] if code == constants.HTTP_FORBIDDEN: raise housecanary.exceptions.UnauthorizedException(code, message) if code == constants.HTTP_TOO_MANY_REQUESTS: raise housecanary.exceptions.RateLimitException(code, message, response) else: raise housecanary.exceptions.RequestException(code, message) request_url = response.request.url endpoint_name = self._parse_endpoint_name_from_url(request_url) return Response.create(endpoint_name, response_json, response)
process_json_response
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, response):\n response_json = response.json()\n code_key = 'code'\n if code_key in response_json and response_json[code_key\n ] != constants.HTTP_CODE_OK:\n code = response_json[code_key]\n message = response_json\n if 'message' in response_json:\n message = response_json['message']\n elif 'code_description' in response_json:\n message = response_json['code_description']\n if code == constants.HTTP_FORBIDDEN:\n raise housecanary.exceptions.UnauthorizedException(code, message)\n if code == constants.HTTP_TOO_MANY_REQUESTS:\n raise housecanary.exceptions.RateLimitException(code, message,\n response)\n else:\n raise housecanary.exceptions.RequestException(code, message)\n request_url = response.request.url\n endpoint_name = self._parse_endpoint_name_from_url(request_url)\n return Response.create(endpoint_name, response_json, response)\n", "role": "user" }, { "content": "process_json_response", "role": "assistant" } ]
def x(self): watcher_thread = threading.Thread(target=self.run_watcher) if self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT: daemon = False else: daemon = True watcher_thread.setDaemon(daemon) watcher_thread.start() return watcher_thread
start_watcher_thread
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n watcher_thread = threading.Thread(target=self.run_watcher)\n if self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT:\n daemon = False\n else:\n daemon = True\n watcher_thread.setDaemon(daemon)\n watcher_thread.start()\n return watcher_thread\n", "role": "user" }, { "content": "start_watcher_thread", "role": "assistant" } ]
def x(self): observer = Observer() observer.start() watche_obj_map = {} while not self._watcher_to_stop: old_watch_path_s = set(watche_obj_map) new_watch_path_s = self._find_watch_paths() for new_watch_path in new_watch_path_s: old_watch_path_s.discard(new_watch_path) if new_watch_path not in watche_obj_map: try: watch_obj = observer.schedule(self, new_watch_path, recursive=True) watche_obj_map[new_watch_path] = watch_obj except OSError: watche_obj_map[new_watch_path] = None for old_watch_path in old_watch_path_s: watch_obj = watche_obj_map.pop(old_watch_path, None) if watch_obj is not None: observer.unschedule(watch_obj) self._watch_paths = new_watch_path_s time.sleep(self._interval)
run_watcher
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n observer = Observer()\n observer.start()\n watche_obj_map = {}\n while not self._watcher_to_stop:\n old_watch_path_s = set(watche_obj_map)\n new_watch_path_s = self._find_watch_paths()\n for new_watch_path in new_watch_path_s:\n old_watch_path_s.discard(new_watch_path)\n if new_watch_path not in watche_obj_map:\n try:\n watch_obj = observer.schedule(self, new_watch_path,\n recursive=True)\n watche_obj_map[new_watch_path] = watch_obj\n except OSError:\n watche_obj_map[new_watch_path] = None\n for old_watch_path in old_watch_path_s:\n watch_obj = watche_obj_map.pop(old_watch_path, None)\n if watch_obj is not None:\n observer.unschedule(watch_obj)\n self._watch_paths = new_watch_path_s\n time.sleep(self._interval)\n", "role": "user" }, { "content": "run_watcher", "role": "assistant" } ]
def x(self): watch_path_s = set(os.path.abspath(x) for x in sys.path) for extra_path in (self._extra_paths or ()): extra_dir_path = os.path.dirname(os.path.abspath(extra_path)) watch_path_s.add(extra_dir_path) for module in list(sys.modules.values()): module_path = getattr(module, '__file__', None) if module_path is not None: module_dir_path = os.path.dirname(os.path.abspath(module_path)) watch_path_s.add(module_dir_path) watch_path_s = self._find_short_paths(watch_path_s) return watch_path_s
find_watch_paths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n watch_path_s = set(os.path.abspath(x) for x in sys.path)\n for extra_path in (self._extra_paths or ()):\n extra_dir_path = os.path.dirname(os.path.abspath(extra_path))\n watch_path_s.add(extra_dir_path)\n for module in list(sys.modules.values()):\n module_path = getattr(module, '__file__', None)\n if module_path is not None:\n module_dir_path = os.path.dirname(os.path.abspath(module_path))\n watch_path_s.add(module_dir_path)\n watch_path_s = self._find_short_paths(watch_path_s)\n return watch_path_s\n", "role": "user" }, { "content": "find_watch_paths", "role": "assistant" } ]
def x(self, paths): path_parts_s = [path.split(os.path.sep) for path in paths] root_node = {} for parts in sorted(path_parts_s, key=len, reverse=True): node = root_node for part in parts: node = node.setdefault(part, {}) node.clear() short_path_s = set() self._collect_leaf_paths(node=root_node, path_parts=(), leaf_paths= short_path_s) return short_path_s
find_short_paths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, paths):\n path_parts_s = [path.split(os.path.sep) for path in paths]\n root_node = {}\n for parts in sorted(path_parts_s, key=len, reverse=True):\n node = root_node\n for part in parts:\n node = node.setdefault(part, {})\n node.clear()\n short_path_s = set()\n self._collect_leaf_paths(node=root_node, path_parts=(), leaf_paths=\n short_path_s)\n return short_path_s\n", "role": "user" }, { "content": "find_short_paths", "role": "assistant" } ]
def x(self, node, path_parts, leaf_paths): if not node: node_path = '/'.join(path_parts) leaf_paths.add(node_path) else: for child_path_part, child_node in node.items(): child_path_part_s = path_parts + (child_path_part,) self._collect_leaf_paths(node=child_node, path_parts= child_path_part_s, leaf_paths=leaf_paths)
collect_leaf_paths
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, node, path_parts, leaf_paths):\n if not node:\n node_path = '/'.join(path_parts)\n leaf_paths.add(node_path)\n else:\n for child_path_part, child_node in node.items():\n child_path_part_s = path_parts + (child_path_part,)\n self._collect_leaf_paths(node=child_node, path_parts=\n child_path_part_s, leaf_paths=leaf_paths)\n", "role": "user" }, { "content": "collect_leaf_paths", "role": "assistant" } ]
def x(self, event): file_path = event.src_path if file_path in self._extra_paths: self.reload() if file_path.endswith(('.pyc', '.pyo')): file_path = file_path[:-1] if file_path.endswith('.py'): file_dir = os.path.dirname(file_path) if file_dir.startswith(tuple(self._watch_paths)): self.reload()
dispatch
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self, event):\n file_path = event.src_path\n if file_path in self._extra_paths:\n self.reload()\n if file_path.endswith(('.pyc', '.pyo')):\n file_path = file_path[:-1]\n if file_path.endswith('.py'):\n file_dir = os.path.dirname(file_path)\n if file_dir.startswith(tuple(self._watch_paths)):\n self.reload()\n", "role": "user" }, { "content": "dispatch", "role": "assistant" } ]
def x(self): reload_mode = self._reload_mode if self._reload_mode == self.RELOAD_MODE_V_EXEC: self.reload_using_exec() elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_EXIT: self.reload_using_spawn_exit() elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT: self.reload_using_spawn_wait() else: error_msg = 'Invalid reload mode: {}.'.format(repr(reload_mode)) raise ValueError(error_msg)
reload
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n reload_mode = self._reload_mode\n if self._reload_mode == self.RELOAD_MODE_V_EXEC:\n self.reload_using_exec()\n elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_EXIT:\n self.reload_using_spawn_exit()\n elif self._reload_mode == self.RELOAD_MODE_V_SPAWN_WAIT:\n self.reload_using_spawn_wait()\n else:\n error_msg = 'Invalid reload mode: {}.'.format(repr(reload_mode))\n raise ValueError(error_msg)\n", "role": "user" }, { "content": "reload", "role": "assistant" } ]
def x(self): cmd_parts = [sys.executable] + sys.argv env_copy = os.environ.copy() os.execvpe(sys.executable, cmd_parts, env_copy)
reload_using_exec
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cmd_parts = [sys.executable] + sys.argv\n env_copy = os.environ.copy()\n os.execvpe(sys.executable, cmd_parts, env_copy)\n", "role": "user" }, { "content": "reload_using_exec", "role": "assistant" } ]
def x(self): cmd_parts = [sys.executable] + sys.argv env_copy = os.environ.copy() subprocess.Popen(cmd_parts, env=env_copy, close_fds=True) if self._force_exit: os._exit(0) else: interrupt_main() self._watcher_to_stop = True sys.exit(0)
reload_using_spawn_exit
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cmd_parts = [sys.executable] + sys.argv\n env_copy = os.environ.copy()\n subprocess.Popen(cmd_parts, env=env_copy, close_fds=True)\n if self._force_exit:\n os._exit(0)\n else:\n interrupt_main()\n self._watcher_to_stop = True\n sys.exit(0)\n", "role": "user" }, { "content": "reload_using_spawn_exit", "role": "assistant" } ]
def x(self): cmd_parts = [sys.executable] + sys.argv env_copy = os.environ.copy() interrupt_main() subprocess.call(cmd_parts, env=env_copy, close_fds=True) sys.exit(0)
reload_using_spawn_wait
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(self):\n cmd_parts = [sys.executable] + sys.argv\n env_copy = os.environ.copy()\n interrupt_main()\n subprocess.call(cmd_parts, env=env_copy, close_fds=True)\n sys.exit(0)\n", "role": "user" }, { "content": "reload_using_spawn_wait", "role": "assistant" } ]
def x(agent_id): fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), ( 'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'), ('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots', 'available_slots'), ('Occupied slots', 'occupied_slots')] if is_legacy_server(): del fields[9] del fields[6] q = 'query($agent_id:String!) { agent(agent_id:$agent_id) { $fields }}' q = q.replace('$fields', ' '.join(item[1] for item in fields)) v = {'agent_id': agent_id} with Session() as session: try: resp = session.Admin.query(q, v) except Exception as e: print_error(e) sys.exit(1) info = resp['agent'] rows = [] for name, key in fields: if key == 'mem_cur_bytes' and info[key] is not None: info[key] = round(info[key] / 2 ** 20, 1) if key in info: rows.append((name, info[key])) print(tabulate(rows, headers=('Field', 'Value')))
agent
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(agent_id):\n fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), (\n 'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'),\n ('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots',\n 'available_slots'), ('Occupied slots', 'occupied_slots')]\n if is_legacy_server():\n del fields[9]\n del fields[6]\n q = 'query($agent_id:String!) { agent(agent_id:$agent_id) { $fields }}'\n q = q.replace('$fields', ' '.join(item[1] for item in fields))\n v = {'agent_id': agent_id}\n with Session() as session:\n try:\n resp = session.Admin.query(q, v)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n info = resp['agent']\n rows = []\n for name, key in fields:\n if key == 'mem_cur_bytes' and info[key] is not None:\n info[key] = round(info[key] / 2 ** 20, 1)\n if key in info:\n rows.append((name, info[key]))\n print(tabulate(rows, headers=('Field', 'Value')))\n", "role": "user" }, { "content": "agent", "role": "assistant" } ]
def x(status, all): fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), ( 'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'), ('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots', 'available_slots'), ('Occupied slots', 'occupied_slots')] if is_legacy_server(): del fields[9] del fields[6] def x(limit, offset): try: resp_agents = session.Agent.list_with_limit(limit, offset, status, fields=(item[1] for item in fields)) except Exception as e: print_error(e) sys.exit(1) return resp_agents def x(results): for item in results: if 'mem_cur_bytes' in item and item['mem_cur_bytes'] is not None: item['mem_cur_bytes'] = round(item['mem_cur_bytes'] / 2 ** 20, 1) return results def x(interval): offset = 0 is_first = True total_count = -1 while True: limit = interval if is_first else min(interval, total_count - offset) try: result = execute_paginated_query(limit, offset) except Exception as e: print_error(e) sys.exit(1) offset += interval total_count = result['total_count'] items = result['items'] items = round_mem(items) table = tabulate((item.values() for item in items), headers=( item[0] for item in fields)) if is_first: is_first = False else: table_rows = table.split('\n') table = '\n'.join(table_rows[2:]) yield table + '\n' if not offset < total_count: break with Session() as session: paginating_interval = 10 if all: click.echo_via_pager(_generate_paginated_results( paginating_interval)) else: result = execute_paginated_query(paginating_interval, offset=0) total_count = result['total_count'] if total_count == 0: print('There are no matching agents.') return items = result['items'] items = round_mem(items) fields = [field for field in fields if field[1] in items[0]] print(tabulate((item.values() for item in items), headers=(item [0] for item in fields))) if total_count > paginating_interval: print('More agents can be displayed by using --all option.')
agents
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(status, all):\n fields = [('ID', 'id'), ('Status', 'status'), ('Region', 'region'), (\n 'First Contact', 'first_contact'), ('CPU Usage (%)', 'cpu_cur_pct'),\n ('Used Memory (MiB)', 'mem_cur_bytes'), ('Total slots',\n 'available_slots'), ('Occupied slots', 'occupied_slots')]\n if is_legacy_server():\n del fields[9]\n del fields[6]\n\n def x(limit, offset):\n try:\n resp_agents = session.Agent.list_with_limit(limit, offset,\n status, fields=(item[1] for item in fields))\n except Exception as e:\n print_error(e)\n sys.exit(1)\n return resp_agents\n\n def x(results):\n for item in results:\n if 'mem_cur_bytes' in item and item['mem_cur_bytes'] is not None:\n item['mem_cur_bytes'] = round(item['mem_cur_bytes'] / 2 ** \n 20, 1)\n return results\n\n def x(interval):\n offset = 0\n is_first = True\n total_count = -1\n while True:\n limit = interval if is_first else min(interval, total_count -\n offset)\n try:\n result = execute_paginated_query(limit, offset)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n offset += interval\n total_count = result['total_count']\n items = result['items']\n items = round_mem(items)\n table = tabulate((item.values() for item in items), headers=(\n item[0] for item in fields))\n if is_first:\n is_first = False\n else:\n table_rows = table.split('\\n')\n table = '\\n'.join(table_rows[2:])\n yield table + '\\n'\n if not offset < total_count:\n break\n with Session() as session:\n paginating_interval = 10\n if all:\n click.echo_via_pager(_generate_paginated_results(\n paginating_interval))\n else:\n result = execute_paginated_query(paginating_interval, offset=0)\n total_count = result['total_count']\n if total_count == 0:\n print('There are no matching agents.')\n return\n items = result['items']\n items = round_mem(items)\n fields = [field for field in fields if field[1] in items[0]]\n print(tabulate((item.values() for item in items), headers=(item\n [0] for item in fields)))\n if total_count > paginating_interval:\n print('More agents can be displayed by using --all option.')\n", "role": "user" }, { "content": "agents", "role": "assistant" } ]
def x(name): fields = [('Name', 'name'), ('Created At', 'created_at'), ( 'Default for Unspecified', 'default_for_unspecified'), ( 'Total Resource Slot', 'total_resource_slots'), ( 'Max Concurrent Sessions', 'max_concurrent_sessions'), ( 'Max Containers per Session', 'max_containers_per_session'), ( 'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size', 'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), ( 'Allowed vFolder Hosts', 'allowed_vfolder_hosts')] with Session() as session: try: rp = session.ResourcePolicy(session.config.access_key) info = rp.info(name, fields=(item[1] for item in fields)) except Exception as e: print_error(e) sys.exit(1) rows = [] if info is None: print('No such resource policy.') sys.exit(1) for name, key in fields: rows.append((name, info[key])) print(tabulate(rows, headers=('Field', 'Value')))
resource_policy
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name):\n fields = [('Name', 'name'), ('Created At', 'created_at'), (\n 'Default for Unspecified', 'default_for_unspecified'), (\n 'Total Resource Slot', 'total_resource_slots'), (\n 'Max Concurrent Sessions', 'max_concurrent_sessions'), (\n 'Max Containers per Session', 'max_containers_per_session'), (\n 'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size',\n 'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), (\n 'Allowed vFolder Hosts', 'allowed_vfolder_hosts')]\n with Session() as session:\n try:\n rp = session.ResourcePolicy(session.config.access_key)\n info = rp.info(name, fields=(item[1] for item in fields))\n except Exception as e:\n print_error(e)\n sys.exit(1)\n rows = []\n if info is None:\n print('No such resource policy.')\n sys.exit(1)\n for name, key in fields:\n rows.append((name, info[key]))\n print(tabulate(rows, headers=('Field', 'Value')))\n", "role": "user" }, { "content": "resource_policy", "role": "assistant" } ]
def x(ctx): if ctx.invoked_subcommand is not None: return fields = [('Name', 'name'), ('Created At', 'created_at'), ( 'Default for Unspecified', 'default_for_unspecified'), ( 'Total Resource Slot', 'total_resource_slots'), ( 'Max Concurrent Sessions', 'max_concurrent_sessions'), ( 'Max Containers per Session', 'max_containers_per_session'), ( 'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size', 'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), ( 'Allowed vFolder Hosts', 'allowed_vfolder_hosts')] with Session() as session: try: items = session.ResourcePolicy.list(fields=(item[1] for item in fields)) except Exception as e: print_error(e) sys.exit(1) if len(items) == 0: print('There are no keypair resource policies.') return print(tabulate((item.values() for item in items), headers=(item[0] for item in fields)))
resource_policies
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ctx):\n if ctx.invoked_subcommand is not None:\n return\n fields = [('Name', 'name'), ('Created At', 'created_at'), (\n 'Default for Unspecified', 'default_for_unspecified'), (\n 'Total Resource Slot', 'total_resource_slots'), (\n 'Max Concurrent Sessions', 'max_concurrent_sessions'), (\n 'Max Containers per Session', 'max_containers_per_session'), (\n 'Max vFolder Count', 'max_vfolder_count'), ('Max vFolder Size',\n 'max_vfolder_size'), ('Idle Timeeout', 'idle_timeout'), (\n 'Allowed vFolder Hosts', 'allowed_vfolder_hosts')]\n with Session() as session:\n try:\n items = session.ResourcePolicy.list(fields=(item[1] for item in\n fields))\n except Exception as e:\n print_error(e)\n sys.exit(1)\n if len(items) == 0:\n print('There are no keypair resource policies.')\n return\n print(tabulate((item.values() for item in items), headers=(item[0] for\n item in fields)))\n", "role": "user" }, { "content": "resource_policies", "role": "assistant" } ]
def x(name, default_for_unspecified, total_resource_slots, max_concurrent_sessions, max_containers_per_session, max_vfolder_count, max_vfolder_size, idle_timeout, allowed_vfolder_hosts): with Session() as session: try: data = session.ResourcePolicy.create(name, default_for_unspecified=default_for_unspecified, total_resource_slots=total_resource_slots, max_concurrent_sessions=max_concurrent_sessions, max_containers_per_session=max_containers_per_session, max_vfolder_count=max_vfolder_count, max_vfolder_size= max_vfolder_size, idle_timeout=idle_timeout, allowed_vfolder_hosts=allowed_vfolder_hosts) except Exception as e: print_error(e) sys.exit(1) if not data['ok']: print_fail('KeyPair Resource Policy creation has failed: {0}'. format(data['msg'])) sys.exit(1) item = data['resource_policy'] print('Keypair resource policy ' + item['name'] + ' is created.')
add
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name, default_for_unspecified, total_resource_slots,\n max_concurrent_sessions, max_containers_per_session, max_vfolder_count,\n max_vfolder_size, idle_timeout, allowed_vfolder_hosts):\n with Session() as session:\n try:\n data = session.ResourcePolicy.create(name,\n default_for_unspecified=default_for_unspecified,\n total_resource_slots=total_resource_slots,\n max_concurrent_sessions=max_concurrent_sessions,\n max_containers_per_session=max_containers_per_session,\n max_vfolder_count=max_vfolder_count, max_vfolder_size=\n max_vfolder_size, idle_timeout=idle_timeout,\n allowed_vfolder_hosts=allowed_vfolder_hosts)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n if not data['ok']:\n print_fail('KeyPair Resource Policy creation has failed: {0}'.\n format(data['msg']))\n sys.exit(1)\n item = data['resource_policy']\n print('Keypair resource policy ' + item['name'] + ' is created.')\n", "role": "user" }, { "content": "add", "role": "assistant" } ]
def x(name): with Session() as session: if input('Are you sure? (y/n): ').lower().strip()[:1] != 'y': print('Canceled.') sys.exit(1) try: data = session.ResourcePolicy.delete(name) except Exception as e: print_error(e) sys.exit(1) if not data['ok']: print_fail('KeyPair Resource Policy deletion has failed: {0}'. format(data['msg'])) sys.exit(1) print('Resource policy ' + name + ' is deleted.')
delete
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(name):\n with Session() as session:\n if input('Are you sure? (y/n): ').lower().strip()[:1] != 'y':\n print('Canceled.')\n sys.exit(1)\n try:\n data = session.ResourcePolicy.delete(name)\n except Exception as e:\n print_error(e)\n sys.exit(1)\n if not data['ok']:\n print_fail('KeyPair Resource Policy deletion has failed: {0}'.\n format(data['msg']))\n sys.exit(1)\n print('Resource policy ' + name + ' is deleted.')\n", "role": "user" }, { "content": "delete", "role": "assistant" } ]
def x(session_id, app, bind, port): api_session = None runner = None async def app_setup(): nonlocal api_session, runner loop = current_loop() api_session = AsyncSession() protocol = 'http' runner = ProxyRunner(api_session, session_id, app, protocol, bind, port, loop=loop) await runner.ready() print_info('A local proxy to the application "{0}" '.format(app) + 'provided by the session "{0}" '.format(session_id) + 'is available at: {0}://{1}:{2}'.format(protocol, bind, port)) async def app_shutdown(): nonlocal api_session, runner print_info('Shutting down....') await runner.close() await api_session.close() print_info('The local proxy to "{}" has terminated.'.format(app)) asyncio_run_forever(app_setup(), app_shutdown(), stop_signals={signal. SIGINT, signal.SIGTERM})
app
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(session_id, app, bind, port):\n api_session = None\n runner = None\n\n async def app_setup():\n nonlocal api_session, runner\n loop = current_loop()\n api_session = AsyncSession()\n protocol = 'http'\n runner = ProxyRunner(api_session, session_id, app, protocol, bind,\n port, loop=loop)\n await runner.ready()\n print_info('A local proxy to the application \"{0}\" '.format(app) +\n 'provided by the session \"{0}\" '.format(session_id) +\n 'is available at: {0}://{1}:{2}'.format(protocol, bind, port))\n\n async def app_shutdown():\n nonlocal api_session, runner\n print_info('Shutting down....')\n await runner.close()\n await api_session.close()\n print_info('The local proxy to \"{}\" has terminated.'.format(app))\n asyncio_run_forever(app_setup(), app_shutdown(), stop_signals={signal.\n SIGINT, signal.SIGTERM})\n", "role": "user" }, { "content": "app", "role": "assistant" } ]
def x(key: str, default: Any=None, clean: Callable[[str], Any]=lambda v: v): key = key.upper() v = os.environ.get('BACKEND_' + key) if v is None: v = os.environ.get('SORNA_' + key) if v is None: if default is None: raise KeyError(key) v = default return clean(v)
get_env
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key: str, default: Any=None, clean: Callable[[str], Any]=lambda v: v):\n key = key.upper()\n v = os.environ.get('BACKEND_' + key)\n if v is None:\n v = os.environ.get('SORNA_' + key)\n if v is None:\n if default is None:\n raise KeyError(key)\n v = default\n return clean(v)\n", "role": "user" }, { "content": "get_env", "role": "assistant" } ]
def x(sess_id_or_alias): with Session() as session: try: print_wait('Retrieving container logs...') kernel = session.Kernel(sess_id_or_alias) result = kernel.get_logs().get('result') logs = result.get('logs') if 'logs' in result else '' print(logs) print_done('End of logs.') except Exception as e: print_error(e) sys.exit(1)
logs
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(sess_id_or_alias):\n with Session() as session:\n try:\n print_wait('Retrieving container logs...')\n kernel = session.Kernel(sess_id_or_alias)\n result = kernel.get_logs().get('result')\n logs = result.get('logs') if 'logs' in result else ''\n print(logs)\n print_done('End of logs.')\n except Exception as e:\n print_error(e)\n sys.exit(1)\n", "role": "user" }, { "content": "logs", "role": "assistant" } ]
def x(function, args, kws): return hashlib.md5(pickle.dumps((_from_file(function) + function. __name__, args, kws))).hexdigest()
wrap_key
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(function, args, kws):\n return hashlib.md5(pickle.dumps((_from_file(function) + function.\n __name__, args, kws))).hexdigest()\n", "role": "user" }, { "content": "wrap_key", "role": "assistant" } ]
def x(key, adapter=MemoryAdapter): try: return pickle.loads(adapter().get(key)) except CacheExpiredException: return None
get
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key, adapter=MemoryAdapter):\n try:\n return pickle.loads(adapter().get(key))\n except CacheExpiredException:\n return None\n", "role": "user" }, { "content": "get", "role": "assistant" } ]
def x(key, value, timeout=-1, adapter=MemoryAdapter): if adapter(timeout=timeout).set(key, pickle.dumps(value)): return value else: return None
set
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(key, value, timeout=-1, adapter=MemoryAdapter):\n if adapter(timeout=timeout).set(key, pickle.dumps(value)):\n return value\n else:\n return None\n", "role": "user" }, { "content": "set", "role": "assistant" } ]
def x(timeout=-1, adapter=MemoryAdapter): def x(function): @wraps(function) def x(*args, **kws): hash_key = _wrap_key(function, args, kws) try: adapter_instance = adapter() return pickle.loads(adapter_instance.get(hash_key)) except CacheExpiredException: value = function(*args, **kws) set(hash_key, value, timeout, adapter) return value return __wrapcache return _wrapcache
wrapcache
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(timeout=-1, adapter=MemoryAdapter):\n\n def x(function):\n\n @wraps(function)\n def x(*args, **kws):\n hash_key = _wrap_key(function, args, kws)\n try:\n adapter_instance = adapter()\n return pickle.loads(adapter_instance.get(hash_key))\n except CacheExpiredException:\n value = function(*args, **kws)\n set(hash_key, value, timeout, adapter)\n return value\n return __wrapcache\n return _wrapcache\n", "role": "user" }, { "content": "wrapcache", "role": "assistant" } ]
def x(data, output_file_name, result_info_key, identifier_keys): workbook = create_excel_workbook(data, result_info_key, identifier_keys) workbook.save(output_file_name) print('Saved Excel file to {}'.format(output_file_name))
export_analytics_data_to_excel
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, output_file_name, result_info_key, identifier_keys):\n workbook = create_excel_workbook(data, result_info_key, identifier_keys)\n workbook.save(output_file_name)\n print('Saved Excel file to {}'.format(output_file_name))\n", "role": "user" }, { "content": "export_analytics_data_to_excel", "role": "assistant" } ]
def x(data, output_folder, result_info_key, identifier_keys): workbook = create_excel_workbook(data, result_info_key, identifier_keys) suffix = '.csv' if not os.path.exists(output_folder): os.makedirs(output_folder) for worksheet in workbook.worksheets: file_name = utilities.convert_title_to_snake_case(worksheet.title) file_path = os.path.join(output_folder, file_name + suffix) mode = 'w' if sys.version_info[0] < 3: mode = 'wb' with io.open(file_path, mode) as output_file: csv_writer = csv.writer(output_file) for row in worksheet.rows: csv_writer.writerow([cell.value for cell in row]) print('Saved CSV files to {}'.format(output_folder))
export_analytics_data_to_csv
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, output_folder, result_info_key, identifier_keys):\n workbook = create_excel_workbook(data, result_info_key, identifier_keys)\n suffix = '.csv'\n if not os.path.exists(output_folder):\n os.makedirs(output_folder)\n for worksheet in workbook.worksheets:\n file_name = utilities.convert_title_to_snake_case(worksheet.title)\n file_path = os.path.join(output_folder, file_name + suffix)\n mode = 'w'\n if sys.version_info[0] < 3:\n mode = 'wb'\n with io.open(file_path, mode) as output_file:\n csv_writer = csv.writer(output_file)\n for row in worksheet.rows:\n csv_writer.writerow([cell.value for cell in row])\n print('Saved CSV files to {}'.format(output_folder))\n", "role": "user" }, { "content": "export_analytics_data_to_csv", "role": "assistant" } ]
def x(addresses, output_file_name, endpoint, report_type, retry, api_key, api_secret, files_path): master_workbook = openpyxl.Workbook() if api_key is not None and api_secret is not None: client = ApiClient(api_key, api_secret) else: client = ApiClient() errors = [] for index, addr in enumerate(addresses): print('Processing {}'.format(addr[0])) result = _get_excel_report(client, endpoint, addr[0], addr[1], report_type, retry) if not result['success']: print('Error retrieving report for {}'.format(addr[0])) print(result['content']) errors.append({'address': addr[0], 'message': result['content']}) continue orig_wb = openpyxl.load_workbook(filename=io.BytesIO(result['content']) ) _save_individual_file(orig_wb, files_path, addr[0]) for sheet_name in orig_wb.get_sheet_names(): if sheet_name in master_workbook.get_sheet_names(): master_ws = master_workbook.get_sheet_by_name(sheet_name) else: master_ws = master_workbook.create_sheet(sheet_name) orig_rows = orig_wb.get_sheet_by_name(sheet_name).rows if sheet_name == 'Summary' or sheet_name == 'Chart Data': _process_non_standard_sheet(master_ws, orig_rows, addr, index) continue _process_standard_sheet(master_ws, orig_rows, addr, index) master_workbook.remove(master_workbook.worksheets[0]) if len(errors) > 0: errors_sheet = master_workbook.create_sheet('Errors') for error_idx, error in enumerate(errors): errors_sheet.cell(row=error_idx + 1, column=1, value=error[ 'address']) errors_sheet.cell(row=error_idx + 1, column=2, value=error[ 'message']) adjust_column_width_workbook(master_workbook) output_file_path = os.path.join(files_path, output_file_name) master_workbook.save(output_file_path) print('Saved output to {}'.format(output_file_path))
concat_excel_reports
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(addresses, output_file_name, endpoint, report_type, retry, api_key,\n api_secret, files_path):\n master_workbook = openpyxl.Workbook()\n if api_key is not None and api_secret is not None:\n client = ApiClient(api_key, api_secret)\n else:\n client = ApiClient()\n errors = []\n for index, addr in enumerate(addresses):\n print('Processing {}'.format(addr[0]))\n result = _get_excel_report(client, endpoint, addr[0], addr[1],\n report_type, retry)\n if not result['success']:\n print('Error retrieving report for {}'.format(addr[0]))\n print(result['content'])\n errors.append({'address': addr[0], 'message': result['content']})\n continue\n orig_wb = openpyxl.load_workbook(filename=io.BytesIO(result['content'])\n )\n _save_individual_file(orig_wb, files_path, addr[0])\n for sheet_name in orig_wb.get_sheet_names():\n if sheet_name in master_workbook.get_sheet_names():\n master_ws = master_workbook.get_sheet_by_name(sheet_name)\n else:\n master_ws = master_workbook.create_sheet(sheet_name)\n orig_rows = orig_wb.get_sheet_by_name(sheet_name).rows\n if sheet_name == 'Summary' or sheet_name == 'Chart Data':\n _process_non_standard_sheet(master_ws, orig_rows, addr, index)\n continue\n _process_standard_sheet(master_ws, orig_rows, addr, index)\n master_workbook.remove(master_workbook.worksheets[0])\n if len(errors) > 0:\n errors_sheet = master_workbook.create_sheet('Errors')\n for error_idx, error in enumerate(errors):\n errors_sheet.cell(row=error_idx + 1, column=1, value=error[\n 'address'])\n errors_sheet.cell(row=error_idx + 1, column=2, value=error[\n 'message'])\n adjust_column_width_workbook(master_workbook)\n output_file_path = os.path.join(files_path, output_file_name)\n master_workbook.save(output_file_path)\n print('Saved output to {}'.format(output_file_path))\n", "role": "user" }, { "content": "concat_excel_reports", "role": "assistant" } ]
def x(data, result_info_key, identifier_keys): workbook = analytics_data_excel.get_excel_workbook(data, result_info_key, identifier_keys) adjust_column_width_workbook(workbook) return workbook
create_excel_workbook
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(data, result_info_key, identifier_keys):\n workbook = analytics_data_excel.get_excel_workbook(data,\n result_info_key, identifier_keys)\n adjust_column_width_workbook(workbook)\n return workbook\n", "role": "user" }, { "content": "create_excel_workbook", "role": "assistant" } ]
def x(worksheet): dims = {} padding = 1 for row in worksheet.rows: for cell in row: if not cell.value: continue dims[cell.column] = max(dims.get(cell.column, 0), len(str(cell. value))) for col, value in list(dims.items()): worksheet.column_dimensions[col].width = value + padding
adjust_column_width
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(worksheet):\n dims = {}\n padding = 1\n for row in worksheet.rows:\n for cell in row:\n if not cell.value:\n continue\n dims[cell.column] = max(dims.get(cell.column, 0), len(str(cell.\n value)))\n for col, value in list(dims.items()):\n worksheet.column_dimensions[col].width = value + padding\n", "role": "user" }, { "content": "adjust_column_width", "role": "assistant" } ]
def x(ipaddress, auth, url): get_ap_info_url = ( '/imcrs/wlan/apInfo/queryApBasicInfoByCondition?ipAddress=' + str( ipaddress)) f_url = url + get_ap_info_url payload = None r = requests.get(f_url, auth=auth, headers=HEADERS) try: if r.status_code == 200: if len(r.text) > 0: return json.loads(r.text)['apBasicInfo'] except requests.exceptions.RequestException as e: return 'Error:\n' + str(e) + ' get_ap_info_all: An Error has occured'
get_ap_info
[ { "content": "Given the source code of a python function, suggest a fitting name for the function.", "role": "system" }, { "content": "def x(ipaddress, auth, url):\n get_ap_info_url = (\n '/imcrs/wlan/apInfo/queryApBasicInfoByCondition?ipAddress=' + str(\n ipaddress))\n f_url = url + get_ap_info_url\n payload = None\n r = requests.get(f_url, auth=auth, headers=HEADERS)\n try:\n if r.status_code == 200:\n if len(r.text) > 0:\n return json.loads(r.text)['apBasicInfo']\n except requests.exceptions.RequestException as e:\n return 'Error:\\n' + str(e) + ' get_ap_info_all: An Error has occured'\n", "role": "user" }, { "content": "get_ap_info", "role": "assistant" } ]